var/home/core/zuul-output/0000755000175000017500000000000015114012657014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114021144015462 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004263605415114021135017700 0ustar rootrootDec 03 10:53:42 crc systemd[1]: Starting Kubernetes Kubelet... Dec 03 10:53:42 crc restorecon[4645]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:42 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 10:53:43 crc restorecon[4645]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 10:53:43 crc restorecon[4645]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 03 10:53:43 crc kubenswrapper[4646]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 10:53:43 crc kubenswrapper[4646]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 03 10:53:43 crc kubenswrapper[4646]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 10:53:43 crc kubenswrapper[4646]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 10:53:43 crc kubenswrapper[4646]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 03 10:53:43 crc kubenswrapper[4646]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.662582 4646 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667576 4646 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667605 4646 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667617 4646 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667628 4646 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667637 4646 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667646 4646 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667655 4646 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667664 4646 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667672 4646 feature_gate.go:330] unrecognized feature gate: Example Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667682 4646 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667692 4646 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667714 4646 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667723 4646 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667731 4646 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667740 4646 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667747 4646 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667755 4646 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667763 4646 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667771 4646 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667778 4646 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667786 4646 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667794 4646 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667801 4646 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667811 4646 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667821 4646 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667830 4646 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667839 4646 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667848 4646 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667857 4646 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667865 4646 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667874 4646 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667882 4646 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667890 4646 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667898 4646 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667906 4646 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667914 4646 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667921 4646 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667929 4646 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667938 4646 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667946 4646 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667954 4646 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667965 4646 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667975 4646 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667985 4646 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.667993 4646 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668002 4646 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668009 4646 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668018 4646 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668026 4646 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668034 4646 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668041 4646 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668049 4646 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668056 4646 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668064 4646 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668072 4646 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668080 4646 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668087 4646 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668095 4646 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668103 4646 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668110 4646 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668118 4646 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668126 4646 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668133 4646 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668142 4646 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668149 4646 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668157 4646 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668165 4646 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668172 4646 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668182 4646 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668190 4646 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.668198 4646 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668634 4646 flags.go:64] FLAG: --address="0.0.0.0" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668655 4646 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668670 4646 flags.go:64] FLAG: --anonymous-auth="true" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668682 4646 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668700 4646 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668709 4646 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668721 4646 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668732 4646 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668742 4646 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668751 4646 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668761 4646 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668771 4646 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668783 4646 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668793 4646 flags.go:64] FLAG: --cgroup-root="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668802 4646 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668812 4646 flags.go:64] FLAG: --client-ca-file="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668820 4646 flags.go:64] FLAG: --cloud-config="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668829 4646 flags.go:64] FLAG: --cloud-provider="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668838 4646 flags.go:64] FLAG: --cluster-dns="[]" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668850 4646 flags.go:64] FLAG: --cluster-domain="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668859 4646 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668868 4646 flags.go:64] FLAG: --config-dir="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668877 4646 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668887 4646 flags.go:64] FLAG: --container-log-max-files="5" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668899 4646 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668908 4646 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668917 4646 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668926 4646 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668935 4646 flags.go:64] FLAG: --contention-profiling="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668945 4646 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668954 4646 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668963 4646 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668972 4646 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668983 4646 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.668992 4646 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669001 4646 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669010 4646 flags.go:64] FLAG: --enable-load-reader="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669019 4646 flags.go:64] FLAG: --enable-server="true" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669028 4646 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669039 4646 flags.go:64] FLAG: --event-burst="100" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669049 4646 flags.go:64] FLAG: --event-qps="50" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669058 4646 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669067 4646 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669076 4646 flags.go:64] FLAG: --eviction-hard="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669088 4646 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669096 4646 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669106 4646 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669115 4646 flags.go:64] FLAG: --eviction-soft="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669125 4646 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669134 4646 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669144 4646 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669153 4646 flags.go:64] FLAG: --experimental-mounter-path="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669162 4646 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669170 4646 flags.go:64] FLAG: --fail-swap-on="true" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669179 4646 flags.go:64] FLAG: --feature-gates="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669191 4646 flags.go:64] FLAG: --file-check-frequency="20s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669200 4646 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669210 4646 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669219 4646 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669228 4646 flags.go:64] FLAG: --healthz-port="10248" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669237 4646 flags.go:64] FLAG: --help="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669246 4646 flags.go:64] FLAG: --hostname-override="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669255 4646 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669265 4646 flags.go:64] FLAG: --http-check-frequency="20s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669274 4646 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669283 4646 flags.go:64] FLAG: --image-credential-provider-config="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669291 4646 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669300 4646 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669309 4646 flags.go:64] FLAG: --image-service-endpoint="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669318 4646 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669327 4646 flags.go:64] FLAG: --kube-api-burst="100" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669363 4646 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669373 4646 flags.go:64] FLAG: --kube-api-qps="50" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669382 4646 flags.go:64] FLAG: --kube-reserved="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669391 4646 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669399 4646 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669411 4646 flags.go:64] FLAG: --kubelet-cgroups="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669420 4646 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669428 4646 flags.go:64] FLAG: --lock-file="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669437 4646 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669446 4646 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669456 4646 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669469 4646 flags.go:64] FLAG: --log-json-split-stream="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669478 4646 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669499 4646 flags.go:64] FLAG: --log-text-split-stream="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669508 4646 flags.go:64] FLAG: --logging-format="text" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669517 4646 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669527 4646 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669535 4646 flags.go:64] FLAG: --manifest-url="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669544 4646 flags.go:64] FLAG: --manifest-url-header="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669564 4646 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669574 4646 flags.go:64] FLAG: --max-open-files="1000000" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669585 4646 flags.go:64] FLAG: --max-pods="110" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669594 4646 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669603 4646 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669612 4646 flags.go:64] FLAG: --memory-manager-policy="None" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669621 4646 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669630 4646 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669639 4646 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669648 4646 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669667 4646 flags.go:64] FLAG: --node-status-max-images="50" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669676 4646 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669686 4646 flags.go:64] FLAG: --oom-score-adj="-999" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669695 4646 flags.go:64] FLAG: --pod-cidr="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669703 4646 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669717 4646 flags.go:64] FLAG: --pod-manifest-path="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669726 4646 flags.go:64] FLAG: --pod-max-pids="-1" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669735 4646 flags.go:64] FLAG: --pods-per-core="0" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669745 4646 flags.go:64] FLAG: --port="10250" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669755 4646 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669763 4646 flags.go:64] FLAG: --provider-id="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669772 4646 flags.go:64] FLAG: --qos-reserved="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669782 4646 flags.go:64] FLAG: --read-only-port="10255" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669791 4646 flags.go:64] FLAG: --register-node="true" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669800 4646 flags.go:64] FLAG: --register-schedulable="true" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669808 4646 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669822 4646 flags.go:64] FLAG: --registry-burst="10" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669831 4646 flags.go:64] FLAG: --registry-qps="5" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669840 4646 flags.go:64] FLAG: --reserved-cpus="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669849 4646 flags.go:64] FLAG: --reserved-memory="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669861 4646 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669870 4646 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669880 4646 flags.go:64] FLAG: --rotate-certificates="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669889 4646 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669897 4646 flags.go:64] FLAG: --runonce="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669906 4646 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669915 4646 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669925 4646 flags.go:64] FLAG: --seccomp-default="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669933 4646 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669942 4646 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669952 4646 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669961 4646 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669970 4646 flags.go:64] FLAG: --storage-driver-password="root" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669979 4646 flags.go:64] FLAG: --storage-driver-secure="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669988 4646 flags.go:64] FLAG: --storage-driver-table="stats" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.669997 4646 flags.go:64] FLAG: --storage-driver-user="root" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670005 4646 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670015 4646 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670025 4646 flags.go:64] FLAG: --system-cgroups="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670033 4646 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670048 4646 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670057 4646 flags.go:64] FLAG: --tls-cert-file="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670066 4646 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670078 4646 flags.go:64] FLAG: --tls-min-version="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670087 4646 flags.go:64] FLAG: --tls-private-key-file="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670096 4646 flags.go:64] FLAG: --topology-manager-policy="none" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670105 4646 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670114 4646 flags.go:64] FLAG: --topology-manager-scope="container" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670123 4646 flags.go:64] FLAG: --v="2" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670134 4646 flags.go:64] FLAG: --version="false" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670145 4646 flags.go:64] FLAG: --vmodule="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670156 4646 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.670165 4646 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670393 4646 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670405 4646 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670414 4646 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670424 4646 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670433 4646 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670442 4646 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670450 4646 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670458 4646 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670466 4646 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670474 4646 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670482 4646 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670490 4646 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670497 4646 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670506 4646 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670513 4646 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670521 4646 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670529 4646 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670536 4646 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670544 4646 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670552 4646 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670560 4646 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670574 4646 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670582 4646 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670590 4646 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670598 4646 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670606 4646 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670614 4646 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670622 4646 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670630 4646 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670638 4646 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670646 4646 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670656 4646 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670666 4646 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670675 4646 feature_gate.go:330] unrecognized feature gate: Example Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670683 4646 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670694 4646 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670703 4646 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670713 4646 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670722 4646 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670732 4646 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670741 4646 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670749 4646 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670757 4646 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670765 4646 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670773 4646 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670783 4646 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670792 4646 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670801 4646 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670808 4646 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670816 4646 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670824 4646 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670832 4646 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670840 4646 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670849 4646 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670857 4646 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670864 4646 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670872 4646 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670880 4646 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670887 4646 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670895 4646 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670903 4646 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670911 4646 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670918 4646 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670929 4646 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670937 4646 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670945 4646 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670954 4646 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670961 4646 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670971 4646 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670980 4646 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.670988 4646 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.671066 4646 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.683202 4646 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.683250 4646 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683400 4646 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683419 4646 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683427 4646 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683433 4646 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683439 4646 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683474 4646 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683480 4646 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683486 4646 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683491 4646 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683497 4646 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683503 4646 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683508 4646 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683512 4646 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683518 4646 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683523 4646 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683528 4646 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683532 4646 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683537 4646 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683542 4646 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683548 4646 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683553 4646 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683558 4646 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683563 4646 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683568 4646 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683573 4646 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683578 4646 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683583 4646 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683587 4646 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683592 4646 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683597 4646 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683602 4646 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683608 4646 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683612 4646 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683617 4646 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683623 4646 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683628 4646 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683632 4646 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683638 4646 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683643 4646 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683648 4646 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683653 4646 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683659 4646 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683664 4646 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683670 4646 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683675 4646 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683682 4646 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683690 4646 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683699 4646 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683705 4646 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683711 4646 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683717 4646 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683723 4646 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683728 4646 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683735 4646 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683741 4646 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683746 4646 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683751 4646 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683756 4646 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683761 4646 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683768 4646 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683773 4646 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683779 4646 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683785 4646 feature_gate.go:330] unrecognized feature gate: Example Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683790 4646 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683797 4646 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683804 4646 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683811 4646 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683818 4646 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683824 4646 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683831 4646 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.683837 4646 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.683849 4646 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684016 4646 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684025 4646 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684031 4646 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684037 4646 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684041 4646 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684048 4646 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684055 4646 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684062 4646 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684069 4646 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684076 4646 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684087 4646 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684098 4646 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684105 4646 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684112 4646 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684119 4646 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684123 4646 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684128 4646 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684133 4646 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684138 4646 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684143 4646 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684148 4646 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684153 4646 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684158 4646 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684163 4646 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684168 4646 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684173 4646 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684177 4646 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684182 4646 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684187 4646 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684192 4646 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684197 4646 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684202 4646 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684206 4646 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684212 4646 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684217 4646 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684221 4646 feature_gate.go:330] unrecognized feature gate: Example Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684226 4646 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684232 4646 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684237 4646 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684242 4646 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684249 4646 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684256 4646 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684262 4646 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684268 4646 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684273 4646 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684279 4646 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684284 4646 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684289 4646 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684294 4646 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684299 4646 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684304 4646 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684309 4646 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684314 4646 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684319 4646 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684324 4646 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684329 4646 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684357 4646 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684363 4646 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684368 4646 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684373 4646 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684378 4646 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684384 4646 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684391 4646 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684396 4646 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684402 4646 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684409 4646 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684414 4646 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684419 4646 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684425 4646 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684430 4646 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.684435 4646 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.684443 4646 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.684642 4646 server.go:940] "Client rotation is on, will bootstrap in background" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.687778 4646 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.687885 4646 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.688598 4646 server.go:997] "Starting client certificate rotation" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.688630 4646 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.688845 4646 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-17 10:38:49.402500834 +0000 UTC Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.689050 4646 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 335h45m5.713456163s for next certificate rotation Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.693656 4646 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.695984 4646 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.702659 4646 log.go:25] "Validated CRI v1 runtime API" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.718977 4646 log.go:25] "Validated CRI v1 image API" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.723600 4646 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.727576 4646 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-03-10-48-10-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.727640 4646 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.750321 4646 manager.go:217] Machine: {Timestamp:2025-12-03 10:53:43.748546343 +0000 UTC m=+0.211602568 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199472640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:10008c78-4919-4934-8d65-42e9210ff073 BootID:555f41f0-f8e0-41bc-91f8-8b81b382648c Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599734272 Type:vfs Inodes:3076107 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039894528 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599738368 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076107 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:2e:29:99 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:2e:29:99 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:53:be:58 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:ae:1b:8b Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:23:01:29 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:f0:0f:73 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:2f:45:ba Speed:-1 Mtu:1496} {Name:eth10 MacAddress:2a:f3:84:77:21:1a Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:6e:7a:42:ee:3e:83 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199472640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.751118 4646 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.751560 4646 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.752188 4646 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.752639 4646 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.752725 4646 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.754506 4646 topology_manager.go:138] "Creating topology manager with none policy" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.754631 4646 container_manager_linux.go:303] "Creating device plugin manager" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.754994 4646 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.755119 4646 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.755792 4646 state_mem.go:36] "Initialized new in-memory state store" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.756039 4646 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.758499 4646 kubelet.go:418] "Attempting to sync node with API server" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.758537 4646 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.758568 4646 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.758586 4646 kubelet.go:324] "Adding apiserver pod source" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.758599 4646 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.761255 4646 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.761793 4646 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.762814 4646 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.762983 4646 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.763048 4646 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 03 10:53:43 crc kubenswrapper[4646]: E1203 10:53:43.763138 4646 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 03 10:53:43 crc kubenswrapper[4646]: E1203 10:53:43.763172 4646 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.763847 4646 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.763960 4646 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.764033 4646 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.764129 4646 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.764212 4646 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.764281 4646 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.764365 4646 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.764443 4646 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.764543 4646 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.764620 4646 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.764698 4646 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.764765 4646 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.765137 4646 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.765743 4646 server.go:1280] "Started kubelet" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.766171 4646 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.766836 4646 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.767362 4646 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.768114 4646 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.768704 4646 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.768754 4646 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 03 10:53:43 crc systemd[1]: Started Kubernetes Kubelet. Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.769429 4646 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.769471 4646 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.770171 4646 server.go:460] "Adding debug handlers to kubelet server" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.770584 4646 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.770920 4646 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 08:16:06.922081497 +0000 UTC Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.770995 4646 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 237h22m23.151093044s for next certificate rotation Dec 03 10:53:43 crc kubenswrapper[4646]: E1203 10:53:43.771051 4646 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 10:53:43 crc kubenswrapper[4646]: E1203 10:53:43.771360 4646 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="200ms" Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.771533 4646 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 03 10:53:43 crc kubenswrapper[4646]: E1203 10:53:43.771597 4646 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 03 10:53:43 crc kubenswrapper[4646]: E1203 10:53:43.770127 4646 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.155:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187daf29f6a5f4ad default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 10:53:43.765693613 +0000 UTC m=+0.228749778,LastTimestamp:2025-12-03 10:53:43.765693613 +0000 UTC m=+0.228749778,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.772688 4646 factory.go:55] Registering systemd factory Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.772710 4646 factory.go:221] Registration of the systemd container factory successfully Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.773552 4646 factory.go:153] Registering CRI-O factory Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.773583 4646 factory.go:221] Registration of the crio container factory successfully Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.773683 4646 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.773716 4646 factory.go:103] Registering Raw factory Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.773741 4646 manager.go:1196] Started watching for new ooms in manager Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.774688 4646 manager.go:319] Starting recovery of all containers Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802085 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802178 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802210 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802238 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802267 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802297 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802327 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802410 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802446 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802479 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802511 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802585 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802614 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802651 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802679 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802705 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802750 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802813 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802846 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802873 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802902 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802928 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.802957 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803016 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803044 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803079 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803115 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803145 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803172 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803198 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803223 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803261 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803288 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803360 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803401 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803434 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803461 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803486 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803513 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803539 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803567 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803595 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803623 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803700 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803726 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803753 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803779 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803807 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803837 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803863 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803890 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803920 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.803957 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.804022 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.804058 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.804087 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.804115 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.804143 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.804169 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.804194 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.804222 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.804251 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.804276 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.804301 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.804326 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.809974 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810003 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810030 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810074 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810102 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810129 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810158 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810185 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810218 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810245 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810275 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810299 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810325 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810410 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810438 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810463 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810492 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810520 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810546 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810572 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810598 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810625 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810650 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810677 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810709 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810736 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810762 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810790 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810815 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810843 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810872 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810898 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810925 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.810978 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811008 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811036 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811062 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811090 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811119 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811164 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811191 4646 manager.go:324] Recovery completed Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811196 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811483 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811513 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811527 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811540 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811552 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811564 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811577 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811588 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811598 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811608 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811619 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811629 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811638 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811650 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811659 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811670 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811680 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811690 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811701 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811712 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811722 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811732 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811743 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811754 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811765 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811775 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811785 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811796 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811811 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811821 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811831 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811842 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811852 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811864 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811875 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811886 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811896 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811907 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811922 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.811936 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812637 4646 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812675 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812688 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812698 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812710 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812722 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812733 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812745 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812787 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812798 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812809 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812821 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812832 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812842 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812853 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812866 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812878 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812891 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812902 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812912 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812923 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812934 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812944 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812954 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812965 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812976 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812986 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.812998 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813008 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813018 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813030 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813039 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813050 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813062 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813073 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813083 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813094 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813105 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813115 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813127 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813139 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813153 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813164 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813200 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813210 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813221 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813232 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813243 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813254 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813266 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813276 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813287 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813298 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813310 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813320 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813344 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813356 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813366 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813376 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813387 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813397 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813409 4646 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813419 4646 reconstruct.go:97] "Volume reconstruction finished" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.813425 4646 reconciler.go:26] "Reconciler: start to sync state" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.821589 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.823428 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.823465 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.823477 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.825309 4646 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.825388 4646 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.825429 4646 state_mem.go:36] "Initialized new in-memory state store" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.836441 4646 policy_none.go:49] "None policy: Start" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.841016 4646 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.841042 4646 state_mem.go:35] "Initializing new in-memory state store" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.845327 4646 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.846788 4646 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.846900 4646 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.846996 4646 kubelet.go:2335] "Starting kubelet main sync loop" Dec 03 10:53:43 crc kubenswrapper[4646]: E1203 10:53:43.847101 4646 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 03 10:53:43 crc kubenswrapper[4646]: W1203 10:53:43.848345 4646 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 03 10:53:43 crc kubenswrapper[4646]: E1203 10:53:43.848407 4646 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 03 10:53:43 crc kubenswrapper[4646]: E1203 10:53:43.871945 4646 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.899464 4646 manager.go:334] "Starting Device Plugin manager" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.899536 4646 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.899549 4646 server.go:79] "Starting device plugin registration server" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.900429 4646 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.900447 4646 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.900789 4646 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.900884 4646 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.900899 4646 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 03 10:53:43 crc kubenswrapper[4646]: E1203 10:53:43.906447 4646 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.948212 4646 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.948379 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.949633 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.949679 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.949692 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.949873 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.950093 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.950156 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.951702 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.951767 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.951787 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.952034 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.952835 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.952883 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.953911 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.953946 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.953911 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.953959 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.953971 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.954033 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.954860 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.954885 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.954894 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.955016 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.955135 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.955160 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.956058 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.956186 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.956402 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.956078 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.956565 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.956588 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.956783 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.956835 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.956861 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.957799 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.957925 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.958008 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.958020 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.957979 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.958103 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.958182 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.958204 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.959063 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.959102 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:43 crc kubenswrapper[4646]: I1203 10:53:43.959119 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:43 crc kubenswrapper[4646]: E1203 10:53:43.972227 4646 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="400ms" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.000546 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.001975 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.002057 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.002077 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.002142 4646 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 10:53:44 crc kubenswrapper[4646]: E1203 10:53:44.003034 4646 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.155:6443: connect: connection refused" node="crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.016552 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.016591 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.016612 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.016630 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.016646 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.016660 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.016675 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.016691 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.016711 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.016727 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.016768 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.016881 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.016927 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.016966 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.017022 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.117940 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118013 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118047 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118081 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118116 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118149 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118180 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118186 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118213 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118233 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118252 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118302 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118245 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118277 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118300 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118296 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118315 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118185 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118481 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118554 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118573 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118624 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118645 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118659 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118686 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118720 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118755 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118812 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118854 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.118992 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.204254 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.206624 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.206682 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.206702 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.206745 4646 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 10:53:44 crc kubenswrapper[4646]: E1203 10:53:44.207317 4646 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.155:6443: connect: connection refused" node="crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.304699 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.315115 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.334904 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: W1203 10:53:44.340501 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-3fa68aead75d9d83d311c0fd42a35a6e9cf4f6b91d589b0b9eef3b1a4f9f5f58 WatchSource:0}: Error finding container 3fa68aead75d9d83d311c0fd42a35a6e9cf4f6b91d589b0b9eef3b1a4f9f5f58: Status 404 returned error can't find the container with id 3fa68aead75d9d83d311c0fd42a35a6e9cf4f6b91d589b0b9eef3b1a4f9f5f58 Dec 03 10:53:44 crc kubenswrapper[4646]: W1203 10:53:44.343135 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-29395396f1272579ec2994dc4ecce201c61315c35407f2f3d4d96a2bcf528117 WatchSource:0}: Error finding container 29395396f1272579ec2994dc4ecce201c61315c35407f2f3d4d96a2bcf528117: Status 404 returned error can't find the container with id 29395396f1272579ec2994dc4ecce201c61315c35407f2f3d4d96a2bcf528117 Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.356944 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: W1203 10:53:44.358594 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-e1c44a189dfbb1ef92328283f06a7059c7b69ff073cff335351b0f18ed48c753 WatchSource:0}: Error finding container e1c44a189dfbb1ef92328283f06a7059c7b69ff073cff335351b0f18ed48c753: Status 404 returned error can't find the container with id e1c44a189dfbb1ef92328283f06a7059c7b69ff073cff335351b0f18ed48c753 Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.367034 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 10:53:44 crc kubenswrapper[4646]: E1203 10:53:44.372955 4646 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="800ms" Dec 03 10:53:44 crc kubenswrapper[4646]: W1203 10:53:44.389814 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-89382c64b434a613bc6a8c9e04bac5172f0fac2d2c0b09fae9f7d59fdf794f83 WatchSource:0}: Error finding container 89382c64b434a613bc6a8c9e04bac5172f0fac2d2c0b09fae9f7d59fdf794f83: Status 404 returned error can't find the container with id 89382c64b434a613bc6a8c9e04bac5172f0fac2d2c0b09fae9f7d59fdf794f83 Dec 03 10:53:44 crc kubenswrapper[4646]: W1203 10:53:44.399927 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-5f0090290a2e7e51945ddfc6116a235a850512a29fddc5a5a0376791d21d97fe WatchSource:0}: Error finding container 5f0090290a2e7e51945ddfc6116a235a850512a29fddc5a5a0376791d21d97fe: Status 404 returned error can't find the container with id 5f0090290a2e7e51945ddfc6116a235a850512a29fddc5a5a0376791d21d97fe Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.608056 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.609535 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.609573 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.609583 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.609610 4646 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 10:53:44 crc kubenswrapper[4646]: E1203 10:53:44.610069 4646 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.155:6443: connect: connection refused" node="crc" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.767476 4646 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 03 10:53:44 crc kubenswrapper[4646]: W1203 10:53:44.769846 4646 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 03 10:53:44 crc kubenswrapper[4646]: E1203 10:53:44.769896 4646 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 03 10:53:44 crc kubenswrapper[4646]: W1203 10:53:44.781412 4646 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 03 10:53:44 crc kubenswrapper[4646]: E1203 10:53:44.781458 4646 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.853407 4646 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="351da97d4bd82626edfa05e5ad62684b5f7ed5b8c5b3212335ccc8a48fc5ee6c" exitCode=0 Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.853505 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"351da97d4bd82626edfa05e5ad62684b5f7ed5b8c5b3212335ccc8a48fc5ee6c"} Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.853675 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"89382c64b434a613bc6a8c9e04bac5172f0fac2d2c0b09fae9f7d59fdf794f83"} Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.853787 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.855000 4646 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6" exitCode=0 Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.855069 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6"} Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.855115 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e1c44a189dfbb1ef92328283f06a7059c7b69ff073cff335351b0f18ed48c753"} Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.855223 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.855445 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.855490 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.855501 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.856632 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.856655 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.856665 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.858114 4646 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0" exitCode=0 Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.858171 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0"} Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.858199 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3fa68aead75d9d83d311c0fd42a35a6e9cf4f6b91d589b0b9eef3b1a4f9f5f58"} Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.858300 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.859115 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.859139 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.859149 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.861095 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.862674 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.862689 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.862697 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.863788 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e"} Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.863819 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"29395396f1272579ec2994dc4ecce201c61315c35407f2f3d4d96a2bcf528117"} Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.870780 4646 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ca2a854ac65ad7ca89578287487b425eb23e22940b724e1b15cf94aea70a16d4" exitCode=0 Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.870821 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ca2a854ac65ad7ca89578287487b425eb23e22940b724e1b15cf94aea70a16d4"} Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.870865 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5f0090290a2e7e51945ddfc6116a235a850512a29fddc5a5a0376791d21d97fe"} Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.871025 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.872454 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.872496 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:44 crc kubenswrapper[4646]: I1203 10:53:44.872506 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:44 crc kubenswrapper[4646]: W1203 10:53:44.960934 4646 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 03 10:53:44 crc kubenswrapper[4646]: E1203 10:53:44.961014 4646 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 03 10:53:44 crc kubenswrapper[4646]: W1203 10:53:44.977902 4646 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.155:6443: connect: connection refused Dec 03 10:53:44 crc kubenswrapper[4646]: E1203 10:53:44.977976 4646 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.155:6443: connect: connection refused" logger="UnhandledError" Dec 03 10:53:45 crc kubenswrapper[4646]: E1203 10:53:45.174589 4646 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="1.6s" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.412486 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.417807 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.417864 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.417876 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.417902 4646 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 10:53:45 crc kubenswrapper[4646]: E1203 10:53:45.422402 4646 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.155:6443: connect: connection refused" node="crc" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.877484 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4"} Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.877520 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f"} Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.877531 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0"} Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.877604 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.878658 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.878679 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.878687 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.880121 4646 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="77b4730a7eac9d5b2918aa054bfcd828ecfe96d194767d693d06250bc9563006" exitCode=0 Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.880157 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"77b4730a7eac9d5b2918aa054bfcd828ecfe96d194767d693d06250bc9563006"} Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.880219 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.880728 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.880743 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.880751 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.883196 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"b7b21cc2d14253df86c1188cca5f1d7a15dc6b4b938f6c7f71a6772870d6b659"} Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.883351 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.884197 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.884222 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.884252 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.887449 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a30f13d59d6c9c0046b62a4e9b202ba22e20014327ebd11dc59dc871a318826c"} Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.887500 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"821dd6864300c8286093c2ac960632ba03ed444f90994aa9cc24080e93472a5b"} Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.887513 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c8cbb08ab6f4da24ce8dcbafd255b9dd3461bc80abbd6d750e34abcbac41bb71"} Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.887644 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.888391 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.888441 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.888452 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.890689 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde"} Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.890711 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943"} Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.890721 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d"} Dec 03 10:53:45 crc kubenswrapper[4646]: I1203 10:53:45.890730 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd"} Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.897952 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d"} Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.898134 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.899164 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.899207 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.899229 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.901896 4646 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="915b860bb25c7af21fd6e6cdc418edc7803f1d5fa758ed720f1456acdf90f019" exitCode=0 Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.901985 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"915b860bb25c7af21fd6e6cdc418edc7803f1d5fa758ed720f1456acdf90f019"} Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.902036 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.902238 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.903162 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.903185 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.903193 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.903846 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.903874 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.905649 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:46 crc kubenswrapper[4646]: I1203 10:53:46.979967 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.023177 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.024316 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.024376 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.024389 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.024416 4646 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.908691 4646 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.908734 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.909093 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ce713ed7aefbc61a0dab6ca61e8b6b89f823a705d47e4762aca275b9655ebc37"} Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.909119 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c4d818924c59f4b56774f1f5233d9535c49bafbbc7e8d768e528ffb1a928d98e"} Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.909130 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f0e39245bf4f178b0fd60594327ba12e3e8be1da5b5d2a8fd8ece07cd7573ef7"} Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.909139 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"229aa4d192642d996523cec286f8295b5b1852f2e60432f0a4f30c9e0e68e2f6"} Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.909147 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fa5ee1673d1e461a29768e8f36d119da6cb7482c5af6ac2895b08671c7ef9db0"} Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.909459 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.909854 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.910484 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.910511 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.910523 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.911220 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.911242 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.911251 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.911653 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.911725 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:47 crc kubenswrapper[4646]: I1203 10:53:47.911753 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:48 crc kubenswrapper[4646]: I1203 10:53:48.594640 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:53:48 crc kubenswrapper[4646]: I1203 10:53:48.605007 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:53:48 crc kubenswrapper[4646]: I1203 10:53:48.610666 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:48 crc kubenswrapper[4646]: I1203 10:53:48.910913 4646 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 10:53:48 crc kubenswrapper[4646]: I1203 10:53:48.910961 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:48 crc kubenswrapper[4646]: I1203 10:53:48.911011 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:48 crc kubenswrapper[4646]: I1203 10:53:48.912097 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:48 crc kubenswrapper[4646]: I1203 10:53:48.912140 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:48 crc kubenswrapper[4646]: I1203 10:53:48.912156 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:48 crc kubenswrapper[4646]: I1203 10:53:48.912547 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:48 crc kubenswrapper[4646]: I1203 10:53:48.912599 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:48 crc kubenswrapper[4646]: I1203 10:53:48.912619 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:49 crc kubenswrapper[4646]: I1203 10:53:49.914641 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:49 crc kubenswrapper[4646]: I1203 10:53:49.916403 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:49 crc kubenswrapper[4646]: I1203 10:53:49.916468 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:49 crc kubenswrapper[4646]: I1203 10:53:49.916486 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.316428 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.496652 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.496827 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.498091 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.498121 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.498132 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.663681 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.664070 4646 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.664134 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.665767 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.665827 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.665845 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.779538 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.917092 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.918666 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.918716 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:50 crc kubenswrapper[4646]: I1203 10:53:50.918739 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:51 crc kubenswrapper[4646]: I1203 10:53:51.918858 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:51 crc kubenswrapper[4646]: I1203 10:53:51.919573 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:51 crc kubenswrapper[4646]: I1203 10:53:51.919604 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:51 crc kubenswrapper[4646]: I1203 10:53:51.919613 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:52 crc kubenswrapper[4646]: I1203 10:53:52.366113 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 03 10:53:52 crc kubenswrapper[4646]: I1203 10:53:52.366424 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:52 crc kubenswrapper[4646]: I1203 10:53:52.368397 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:52 crc kubenswrapper[4646]: I1203 10:53:52.368458 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:52 crc kubenswrapper[4646]: I1203 10:53:52.368481 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:53 crc kubenswrapper[4646]: I1203 10:53:53.316745 4646 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 10:53:53 crc kubenswrapper[4646]: I1203 10:53:53.316897 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 10:53:53 crc kubenswrapper[4646]: I1203 10:53:53.324937 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:53 crc kubenswrapper[4646]: I1203 10:53:53.325272 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:53 crc kubenswrapper[4646]: I1203 10:53:53.326712 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:53 crc kubenswrapper[4646]: I1203 10:53:53.326801 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:53 crc kubenswrapper[4646]: I1203 10:53:53.326863 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:53 crc kubenswrapper[4646]: I1203 10:53:53.340006 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 10:53:53 crc kubenswrapper[4646]: I1203 10:53:53.340318 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:53 crc kubenswrapper[4646]: I1203 10:53:53.341811 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:53 crc kubenswrapper[4646]: I1203 10:53:53.342010 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:53 crc kubenswrapper[4646]: I1203 10:53:53.342168 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:53 crc kubenswrapper[4646]: E1203 10:53:53.906799 4646 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 10:53:55 crc kubenswrapper[4646]: I1203 10:53:55.767586 4646 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 03 10:53:56 crc kubenswrapper[4646]: I1203 10:53:56.205909 4646 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 10:53:56 crc kubenswrapper[4646]: I1203 10:53:56.206027 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 10:53:56 crc kubenswrapper[4646]: I1203 10:53:56.220312 4646 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 10:53:56 crc kubenswrapper[4646]: I1203 10:53:56.220408 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 10:53:56 crc kubenswrapper[4646]: I1203 10:53:56.986138 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:53:56 crc kubenswrapper[4646]: I1203 10:53:56.986324 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:56 crc kubenswrapper[4646]: I1203 10:53:56.987685 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:56 crc kubenswrapper[4646]: I1203 10:53:56.987739 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:56 crc kubenswrapper[4646]: I1203 10:53:56.987751 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:57 crc kubenswrapper[4646]: I1203 10:53:57.156856 4646 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 10:53:57 crc kubenswrapper[4646]: I1203 10:53:57.156912 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 10:53:58 crc kubenswrapper[4646]: I1203 10:53:58.623522 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:58 crc kubenswrapper[4646]: I1203 10:53:58.624761 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:58 crc kubenswrapper[4646]: I1203 10:53:58.625169 4646 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 10:53:58 crc kubenswrapper[4646]: I1203 10:53:58.625238 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 10:53:58 crc kubenswrapper[4646]: I1203 10:53:58.626307 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:58 crc kubenswrapper[4646]: I1203 10:53:58.626394 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:58 crc kubenswrapper[4646]: I1203 10:53:58.626407 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:53:58 crc kubenswrapper[4646]: I1203 10:53:58.631448 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:53:58 crc kubenswrapper[4646]: I1203 10:53:58.938293 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:53:58 crc kubenswrapper[4646]: I1203 10:53:58.938998 4646 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 10:53:58 crc kubenswrapper[4646]: I1203 10:53:58.939094 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 10:53:58 crc kubenswrapper[4646]: I1203 10:53:58.939322 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:53:58 crc kubenswrapper[4646]: I1203 10:53:58.939434 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:53:58 crc kubenswrapper[4646]: I1203 10:53:58.939455 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:00 crc kubenswrapper[4646]: I1203 10:54:00.523839 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 03 10:54:00 crc kubenswrapper[4646]: I1203 10:54:00.524218 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:54:00 crc kubenswrapper[4646]: I1203 10:54:00.526098 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:00 crc kubenswrapper[4646]: I1203 10:54:00.526145 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:00 crc kubenswrapper[4646]: I1203 10:54:00.526155 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:00 crc kubenswrapper[4646]: I1203 10:54:00.540970 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 03 10:54:00 crc kubenswrapper[4646]: I1203 10:54:00.942612 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:54:00 crc kubenswrapper[4646]: I1203 10:54:00.943741 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:00 crc kubenswrapper[4646]: I1203 10:54:00.943839 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:00 crc kubenswrapper[4646]: I1203 10:54:00.943901 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.182668 4646 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.184487 4646 trace.go:236] Trace[64895158]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 10:53:48.072) (total time: 13111ms): Dec 03 10:54:01 crc kubenswrapper[4646]: Trace[64895158]: ---"Objects listed" error: 13111ms (10:54:01.184) Dec 03 10:54:01 crc kubenswrapper[4646]: Trace[64895158]: [13.111589424s] [13.111589424s] END Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.184523 4646 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.184781 4646 trace.go:236] Trace[775991984]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 10:53:47.798) (total time: 13385ms): Dec 03 10:54:01 crc kubenswrapper[4646]: Trace[775991984]: ---"Objects listed" error: 13385ms (10:54:01.184) Dec 03 10:54:01 crc kubenswrapper[4646]: Trace[775991984]: [13.385871403s] [13.385871403s] END Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.184920 4646 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.187041 4646 trace.go:236] Trace[1833441853]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 10:53:46.866) (total time: 14320ms): Dec 03 10:54:01 crc kubenswrapper[4646]: Trace[1833441853]: ---"Objects listed" error: 14320ms (10:54:01.186) Dec 03 10:54:01 crc kubenswrapper[4646]: Trace[1833441853]: [14.320508598s] [14.320508598s] END Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.187070 4646 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.189548 4646 trace.go:236] Trace[1654210473]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 10:53:47.407) (total time: 13781ms): Dec 03 10:54:01 crc kubenswrapper[4646]: Trace[1654210473]: ---"Objects listed" error: 13781ms (10:54:01.189) Dec 03 10:54:01 crc kubenswrapper[4646]: Trace[1654210473]: [13.781517027s] [13.781517027s] END Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.189566 4646 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.189609 4646 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.191506 4646 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.254254 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.261417 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.771080 4646 apiserver.go:52] "Watching apiserver" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.773534 4646 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.773746 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-ffm54"] Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.774025 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.774082 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.774107 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.774288 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.774257 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.774366 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.774545 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.774552 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.774667 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ffm54" Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.774811 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.776195 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.776448 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.776682 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.785529 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.785548 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.785572 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.785552 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.785675 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.785529 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.785759 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.787588 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.794857 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.815942 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.826230 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.839473 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.852320 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.859746 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.871736 4646 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.871948 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.881977 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895087 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895467 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895514 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895532 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895553 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895571 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895590 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895605 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895629 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895650 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895668 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895690 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895733 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895756 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895777 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895783 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895797 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895797 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895868 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895897 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895920 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895945 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895967 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895988 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896011 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896034 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896056 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896079 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896104 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896127 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896148 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896167 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896189 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896211 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896230 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896254 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896274 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896299 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896319 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896363 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896392 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896414 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896436 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896458 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896481 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896506 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896527 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896550 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896571 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896622 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896646 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896666 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896687 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896709 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896731 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896756 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896779 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896799 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896820 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896840 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896859 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896880 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896900 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896921 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896943 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896964 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897012 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897033 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897053 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897073 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897097 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897119 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897141 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897165 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897190 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897211 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897234 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897258 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897280 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897302 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897323 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897384 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897410 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897433 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897454 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897477 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897500 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897523 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897543 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897564 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897584 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897606 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897627 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897651 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897674 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897696 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897732 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897758 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897780 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897803 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897824 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897847 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897868 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897890 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897912 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897933 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897953 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897976 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897997 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898020 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898040 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898061 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898082 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898104 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898126 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898149 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898172 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898188 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898207 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898230 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898253 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898275 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898299 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898322 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898363 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898385 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898406 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898423 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898441 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898464 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898485 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898501 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898517 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898538 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898556 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898575 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898592 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898609 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898626 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898643 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898659 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898681 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898696 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898712 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898812 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898832 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898854 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898877 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898901 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898919 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898935 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898953 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898969 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898984 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899001 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899019 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899035 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899051 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899066 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899082 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899098 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899113 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899129 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899145 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899162 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899181 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899198 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899215 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899232 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899248 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899264 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899282 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899299 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899316 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899425 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899447 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899463 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899479 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899496 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899511 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899527 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899544 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899560 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899580 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899603 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899624 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899641 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899659 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899675 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899691 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899707 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899724 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899744 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899770 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899787 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899804 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899847 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899869 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899927 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899953 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899974 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/82e5903b-9412-4fe7-b248-4ad185229bbd-hosts-file\") pod \"node-resolver-ffm54\" (UID: \"82e5903b-9412-4fe7-b248-4ad185229bbd\") " pod="openshift-dns/node-resolver-ffm54" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899993 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900035 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900052 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900071 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900091 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900109 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900126 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900144 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900162 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900181 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4pgc\" (UniqueName: \"kubernetes.io/projected/82e5903b-9412-4fe7-b248-4ad185229bbd-kube-api-access-x4pgc\") pod \"node-resolver-ffm54\" (UID: \"82e5903b-9412-4fe7-b248-4ad185229bbd\") " pod="openshift-dns/node-resolver-ffm54" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900201 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900218 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900238 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900288 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900299 4646 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900922 4646 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.923664 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.924797 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895973 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.895993 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896165 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896358 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896444 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896455 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896585 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896593 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896725 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896737 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896848 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896886 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.896989 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897077 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897109 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897201 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897228 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897243 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897391 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897441 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897650 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897812 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.897837 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898151 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898248 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898285 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898379 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898544 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898712 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898791 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.898946 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899087 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899155 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899365 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899582 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899598 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899749 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.899906 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900081 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900259 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900300 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900481 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900705 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.900912 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.901089 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.902184 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.902680 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.902763 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.903002 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.903087 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.903965 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.904388 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.904938 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.905168 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.909807 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.909935 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.909876 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.910097 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.910069 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.910167 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.910230 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.910243 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.910468 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.910661 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.910788 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.910968 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.911484 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.911687 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.911855 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.913146 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.913194 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.913358 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.913488 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.913731 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.914122 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.914348 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.914492 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.914731 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.915020 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.915260 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.915684 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.915899 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.915922 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.915918 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.915957 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.916047 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.916536 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.916636 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.916662 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.917581 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.917632 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.917770 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.917830 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.917895 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.917987 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.918062 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.918189 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.918368 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.918366 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.918420 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.918553 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.918566 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.918612 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.918631 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.918745 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.918703 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.918907 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.919198 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.919249 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.919301 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.919464 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.919522 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.919631 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.923270 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.923452 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.923683 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.923596 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.923752 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.923934 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.923983 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.923995 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.924014 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.924050 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.924189 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.924526 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.924931 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.924955 4646 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.936203 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:02.436172557 +0000 UTC m=+18.899228692 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.937569 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.937766 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.939850 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.940096 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.940276 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.940455 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.940571 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.925541 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.925566 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.925984 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.926207 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.926240 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.926567 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.926574 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.927262 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.927432 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.927103 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.927788 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.928524 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.928761 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.930186 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.930267 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.930796 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.930863 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.930918 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.930917 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.931239 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.931470 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.931654 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.931743 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.931859 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.931969 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.932123 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.932446 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.932500 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.932616 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.932931 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.933302 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.933664 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.933809 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.934835 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.941350 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.941602 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.941760 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.942217 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.934647 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.953419 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.951813 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.942375 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.944490 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.945027 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.945156 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.946428 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.946979 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.947041 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.947235 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.947579 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.947743 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.947768 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.947975 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.949934 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.948043 4646 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.950196 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.951145 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:54:02.451118821 +0000 UTC m=+18.914174956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.953255 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.955423 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.955993 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.956017 4646 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.955427 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.955645 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.956105 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.956122 4646 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.956443 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.956613 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:02.456587939 +0000 UTC m=+18.919644074 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.956766 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:02.456735214 +0000 UTC m=+18.919791359 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.956896 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:02.456867067 +0000 UTC m=+18.919923202 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.958048 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.963500 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 10:54:01 crc kubenswrapper[4646]: E1203 10:54:01.964090 4646 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.964251 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.965357 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.966902 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.970497 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.974404 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:54:01 crc kubenswrapper[4646]: I1203 10:54:01.980705 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001314 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001383 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001408 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4pgc\" (UniqueName: \"kubernetes.io/projected/82e5903b-9412-4fe7-b248-4ad185229bbd-kube-api-access-x4pgc\") pod \"node-resolver-ffm54\" (UID: \"82e5903b-9412-4fe7-b248-4ad185229bbd\") " pod="openshift-dns/node-resolver-ffm54" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001444 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/82e5903b-9412-4fe7-b248-4ad185229bbd-hosts-file\") pod \"node-resolver-ffm54\" (UID: \"82e5903b-9412-4fe7-b248-4ad185229bbd\") " pod="openshift-dns/node-resolver-ffm54" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001489 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001499 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001508 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001516 4646 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001525 4646 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001535 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001547 4646 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001556 4646 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001564 4646 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001573 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001581 4646 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001591 4646 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001598 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001606 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001615 4646 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001623 4646 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001631 4646 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001639 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001647 4646 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001655 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001663 4646 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001671 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001678 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001686 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001694 4646 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001702 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001711 4646 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001719 4646 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001727 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001735 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001743 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001751 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001759 4646 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001767 4646 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001776 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001785 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001794 4646 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001803 4646 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001811 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001820 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001828 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001836 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001844 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001852 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001860 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001869 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001877 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001886 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001895 4646 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001904 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001911 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001919 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001927 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001936 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001943 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001953 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001962 4646 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001970 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001978 4646 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001988 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.001995 4646 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002003 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002013 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002021 4646 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002028 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002038 4646 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002046 4646 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002054 4646 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002063 4646 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002071 4646 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002078 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002087 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002095 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002103 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002112 4646 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002120 4646 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002129 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002137 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002144 4646 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002152 4646 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002161 4646 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002169 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002176 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002185 4646 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002193 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002201 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002209 4646 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002217 4646 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002226 4646 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002234 4646 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002242 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002250 4646 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002258 4646 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002266 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002274 4646 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002283 4646 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002290 4646 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002298 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002306 4646 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002313 4646 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002322 4646 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002330 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002359 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002372 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002383 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002392 4646 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002414 4646 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002424 4646 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002433 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002442 4646 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002452 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002463 4646 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002474 4646 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002485 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002495 4646 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002505 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002516 4646 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002526 4646 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002537 4646 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002549 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002557 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002567 4646 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002575 4646 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002583 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002592 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002600 4646 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002608 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002615 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002624 4646 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002633 4646 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002641 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002649 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002659 4646 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002667 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002675 4646 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002683 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002691 4646 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002698 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002706 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002714 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002723 4646 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002731 4646 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002739 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002747 4646 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002755 4646 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002763 4646 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002772 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002780 4646 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002789 4646 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002796 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002804 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002812 4646 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002820 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002828 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002835 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002844 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002853 4646 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002860 4646 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002869 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002877 4646 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002885 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002899 4646 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002907 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002915 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002923 4646 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002930 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002938 4646 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002952 4646 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002960 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002969 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002977 4646 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002985 4646 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.002993 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003001 4646 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003010 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003018 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003027 4646 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003036 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003044 4646 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003052 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003060 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003068 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003075 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003082 4646 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003090 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003097 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003105 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003113 4646 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003124 4646 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003131 4646 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003269 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.003442 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.004029 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/82e5903b-9412-4fe7-b248-4ad185229bbd-hosts-file\") pod \"node-resolver-ffm54\" (UID: \"82e5903b-9412-4fe7-b248-4ad185229bbd\") " pod="openshift-dns/node-resolver-ffm54" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.004534 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.004555 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.019285 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4pgc\" (UniqueName: \"kubernetes.io/projected/82e5903b-9412-4fe7-b248-4ad185229bbd-kube-api-access-x4pgc\") pod \"node-resolver-ffm54\" (UID: \"82e5903b-9412-4fe7-b248-4ad185229bbd\") " pod="openshift-dns/node-resolver-ffm54" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.094117 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.105466 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 10:54:02 crc kubenswrapper[4646]: W1203 10:54:02.113857 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-6768a7b7032b6c689fa8d0e963ad49f17dc9b1dfd6033126663eaf4dfcb5c30b WatchSource:0}: Error finding container 6768a7b7032b6c689fa8d0e963ad49f17dc9b1dfd6033126663eaf4dfcb5c30b: Status 404 returned error can't find the container with id 6768a7b7032b6c689fa8d0e963ad49f17dc9b1dfd6033126663eaf4dfcb5c30b Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.114558 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ffm54" Dec 03 10:54:02 crc kubenswrapper[4646]: W1203 10:54:02.120864 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-69daff6e419cbcdf9abb9cf809d316ca7f5af8cd90fa894d05ca0dd201529e2d WatchSource:0}: Error finding container 69daff6e419cbcdf9abb9cf809d316ca7f5af8cd90fa894d05ca0dd201529e2d: Status 404 returned error can't find the container with id 69daff6e419cbcdf9abb9cf809d316ca7f5af8cd90fa894d05ca0dd201529e2d Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.121672 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 10:54:02 crc kubenswrapper[4646]: W1203 10:54:02.132320 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82e5903b_9412_4fe7_b248_4ad185229bbd.slice/crio-7b7c9f27c786e994a980327d000639b9981e165c48bb6783750fcb556fe63501 WatchSource:0}: Error finding container 7b7c9f27c786e994a980327d000639b9981e165c48bb6783750fcb556fe63501: Status 404 returned error can't find the container with id 7b7c9f27c786e994a980327d000639b9981e165c48bb6783750fcb556fe63501 Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.255034 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-lhrsz"] Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.255311 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-lhrsz" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.258856 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.258987 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.259055 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.259384 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.267100 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.279395 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.292632 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.304920 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.311394 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d4dc8da-e71d-4968-a2d4-1b03a616bac4-host\") pod \"node-ca-lhrsz\" (UID: \"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\") " pod="openshift-image-registry/node-ca-lhrsz" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.311422 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1d4dc8da-e71d-4968-a2d4-1b03a616bac4-serviceca\") pod \"node-ca-lhrsz\" (UID: \"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\") " pod="openshift-image-registry/node-ca-lhrsz" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.311446 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4tjp\" (UniqueName: \"kubernetes.io/projected/1d4dc8da-e71d-4968-a2d4-1b03a616bac4-kube-api-access-q4tjp\") pod \"node-ca-lhrsz\" (UID: \"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\") " pod="openshift-image-registry/node-ca-lhrsz" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.321294 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.332744 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.344122 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.368607 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.385689 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.411774 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4tjp\" (UniqueName: \"kubernetes.io/projected/1d4dc8da-e71d-4968-a2d4-1b03a616bac4-kube-api-access-q4tjp\") pod \"node-ca-lhrsz\" (UID: \"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\") " pod="openshift-image-registry/node-ca-lhrsz" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.411843 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d4dc8da-e71d-4968-a2d4-1b03a616bac4-host\") pod \"node-ca-lhrsz\" (UID: \"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\") " pod="openshift-image-registry/node-ca-lhrsz" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.411894 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1d4dc8da-e71d-4968-a2d4-1b03a616bac4-serviceca\") pod \"node-ca-lhrsz\" (UID: \"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\") " pod="openshift-image-registry/node-ca-lhrsz" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.411976 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d4dc8da-e71d-4968-a2d4-1b03a616bac4-host\") pod \"node-ca-lhrsz\" (UID: \"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\") " pod="openshift-image-registry/node-ca-lhrsz" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.412995 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1d4dc8da-e71d-4968-a2d4-1b03a616bac4-serviceca\") pod \"node-ca-lhrsz\" (UID: \"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\") " pod="openshift-image-registry/node-ca-lhrsz" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.428589 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4tjp\" (UniqueName: \"kubernetes.io/projected/1d4dc8da-e71d-4968-a2d4-1b03a616bac4-kube-api-access-q4tjp\") pod \"node-ca-lhrsz\" (UID: \"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\") " pod="openshift-image-registry/node-ca-lhrsz" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.512849 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.512921 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:02 crc kubenswrapper[4646]: E1203 10:54:02.513000 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:54:03.512971084 +0000 UTC m=+19.976027239 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:54:02 crc kubenswrapper[4646]: E1203 10:54:02.513027 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:54:02 crc kubenswrapper[4646]: E1203 10:54:02.513044 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:54:02 crc kubenswrapper[4646]: E1203 10:54:02.513054 4646 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.513062 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:02 crc kubenswrapper[4646]: E1203 10:54:02.513095 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:03.513082807 +0000 UTC m=+19.976138942 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.513117 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.513142 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:02 crc kubenswrapper[4646]: E1203 10:54:02.513271 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:54:02 crc kubenswrapper[4646]: E1203 10:54:02.513287 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:54:02 crc kubenswrapper[4646]: E1203 10:54:02.513300 4646 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:02 crc kubenswrapper[4646]: E1203 10:54:02.513361 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:03.513351265 +0000 UTC m=+19.976407420 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:02 crc kubenswrapper[4646]: E1203 10:54:02.513394 4646 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:54:02 crc kubenswrapper[4646]: E1203 10:54:02.513409 4646 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:54:02 crc kubenswrapper[4646]: E1203 10:54:02.513427 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:03.513418107 +0000 UTC m=+19.976474242 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:54:02 crc kubenswrapper[4646]: E1203 10:54:02.513441 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:03.513433257 +0000 UTC m=+19.976489402 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.571964 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-lhrsz" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.766515 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-cbp2x"] Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.766847 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.774366 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.776269 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.776503 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.776678 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.777659 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.810592 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:02Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815207 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-var-lib-cni-bin\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815297 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-hostroot\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815327 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzm9q\" (UniqueName: \"kubernetes.io/projected/0b003d65-42ef-4a54-96e8-33428e114a1d-kube-api-access-xzm9q\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815408 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-multus-socket-dir-parent\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815432 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-run-k8s-cni-cncf-io\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815488 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0b003d65-42ef-4a54-96e8-33428e114a1d-multus-daemon-config\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815543 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-run-multus-certs\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815589 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-etc-kubernetes\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815614 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-var-lib-kubelet\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815684 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-multus-cni-dir\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815709 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-run-netns\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815742 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-cnibin\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815763 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-var-lib-cni-multus\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815793 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-system-cni-dir\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815812 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-multus-conf-dir\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815839 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-os-release\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.815859 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0b003d65-42ef-4a54-96e8-33428e114a1d-cni-binary-copy\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.829082 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:02Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.845059 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:02Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.847646 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:02 crc kubenswrapper[4646]: E1203 10:54:02.847788 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.865610 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:02Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.884682 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:02Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.897253 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:02Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.908789 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:02Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916685 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-run-netns\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916727 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-var-lib-cni-multus\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916750 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-cnibin\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916776 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-system-cni-dir\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916791 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-multus-conf-dir\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916791 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-run-netns\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916811 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-os-release\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916824 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0b003d65-42ef-4a54-96e8-33428e114a1d-cni-binary-copy\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916837 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-hostroot\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916844 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-cnibin\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916859 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-var-lib-cni-bin\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916847 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-multus-conf-dir\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916799 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-var-lib-cni-multus\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916875 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzm9q\" (UniqueName: \"kubernetes.io/projected/0b003d65-42ef-4a54-96e8-33428e114a1d-kube-api-access-xzm9q\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916886 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-system-cni-dir\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916904 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-run-multus-certs\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916936 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-multus-socket-dir-parent\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916954 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-run-k8s-cni-cncf-io\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916973 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0b003d65-42ef-4a54-96e8-33428e114a1d-multus-daemon-config\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.916990 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-etc-kubernetes\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.917008 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-var-lib-kubelet\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.917026 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-multus-cni-dir\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.917060 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-hostroot\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.917060 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-os-release\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.917085 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-var-lib-cni-bin\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.917091 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-run-k8s-cni-cncf-io\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.917125 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-etc-kubernetes\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.917251 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-multus-cni-dir\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.917298 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-run-multus-certs\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.917303 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-host-var-lib-kubelet\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.917360 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0b003d65-42ef-4a54-96e8-33428e114a1d-multus-socket-dir-parent\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.917858 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0b003d65-42ef-4a54-96e8-33428e114a1d-cni-binary-copy\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.917945 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0b003d65-42ef-4a54-96e8-33428e114a1d-multus-daemon-config\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.921747 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:02Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.932728 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzm9q\" (UniqueName: \"kubernetes.io/projected/0b003d65-42ef-4a54-96e8-33428e114a1d-kube-api-access-xzm9q\") pod \"multus-cbp2x\" (UID: \"0b003d65-42ef-4a54-96e8-33428e114a1d\") " pod="openshift-multus/multus-cbp2x" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.940137 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:02Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.954059 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:02Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.959045 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ffm54" event={"ID":"82e5903b-9412-4fe7-b248-4ad185229bbd","Type":"ContainerStarted","Data":"06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae"} Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.959097 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ffm54" event={"ID":"82e5903b-9412-4fe7-b248-4ad185229bbd","Type":"ContainerStarted","Data":"7b7c9f27c786e994a980327d000639b9981e165c48bb6783750fcb556fe63501"} Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.960428 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-lhrsz" event={"ID":"1d4dc8da-e71d-4968-a2d4-1b03a616bac4","Type":"ContainerStarted","Data":"e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05"} Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.960471 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-lhrsz" event={"ID":"1d4dc8da-e71d-4968-a2d4-1b03a616bac4","Type":"ContainerStarted","Data":"6abfd4f4a8c7f4171072c72039f21337b92dc3902791000deb98f443b1e6da17"} Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.961280 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"69daff6e419cbcdf9abb9cf809d316ca7f5af8cd90fa894d05ca0dd201529e2d"} Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.962414 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d"} Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.962456 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6768a7b7032b6c689fa8d0e963ad49f17dc9b1dfd6033126663eaf4dfcb5c30b"} Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.963788 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.965322 4646 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d" exitCode=255 Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.965361 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d"} Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.967204 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098"} Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.967245 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929"} Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.967261 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b1ec48c55eedc7f88292f0ecf340c7dbd5229dd374c49e328b656cb22425074f"} Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.974496 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:02Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:02 crc kubenswrapper[4646]: I1203 10:54:02.994805 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:02Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.017046 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.033053 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.043943 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.058803 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.072753 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.077253 4646 scope.go:117] "RemoveContainer" containerID="a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.078867 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.085628 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-cbp2x" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.108931 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.125626 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.139351 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.162733 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.172528 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-pndzt"] Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.172913 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.174319 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-b49xq"] Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.177144 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.177363 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.177550 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.177720 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.177879 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.180820 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9lfpq"] Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.182225 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.183132 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.184841 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.188811 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.188993 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.188987 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.189107 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.189186 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.189299 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.189517 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.189732 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.214099 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221704 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-kubelet\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221734 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221752 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-var-lib-openvswitch\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221768 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-etc-openvswitch\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221783 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fef52874-4a7c-4016-938a-69e88790ab19-os-release\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221797 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fef52874-4a7c-4016-938a-69e88790ab19-tuning-conf-dir\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221810 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-systemd\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221824 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-openvswitch\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221837 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-env-overrides\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221852 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-node-log\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221876 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-run-ovn-kubernetes\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221894 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fef52874-4a7c-4016-938a-69e88790ab19-system-cni-dir\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221908 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-run-netns\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221922 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-cni-netd\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221936 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fef52874-4a7c-4016-938a-69e88790ab19-cni-binary-copy\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221953 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-slash\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221969 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fef52874-4a7c-4016-938a-69e88790ab19-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221982 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-log-socket\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.221994 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovnkube-config\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.222016 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h29bd\" (UniqueName: \"kubernetes.io/projected/fef52874-4a7c-4016-938a-69e88790ab19-kube-api-access-h29bd\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.222033 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/9695d947-fc43-4625-86af-a1a3e177bb27-rootfs\") pod \"machine-config-daemon-pndzt\" (UID: \"9695d947-fc43-4625-86af-a1a3e177bb27\") " pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.222062 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-ovn\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.222075 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-systemd-units\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.222092 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jct4d\" (UniqueName: \"kubernetes.io/projected/8cd1981e-44c4-4251-81d1-c07c9d45258b-kube-api-access-jct4d\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.223204 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9695d947-fc43-4625-86af-a1a3e177bb27-proxy-tls\") pod \"machine-config-daemon-pndzt\" (UID: \"9695d947-fc43-4625-86af-a1a3e177bb27\") " pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.223227 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fef52874-4a7c-4016-938a-69e88790ab19-cnibin\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.223250 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9695d947-fc43-4625-86af-a1a3e177bb27-mcd-auth-proxy-config\") pod \"machine-config-daemon-pndzt\" (UID: \"9695d947-fc43-4625-86af-a1a3e177bb27\") " pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.223265 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gg9w\" (UniqueName: \"kubernetes.io/projected/9695d947-fc43-4625-86af-a1a3e177bb27-kube-api-access-6gg9w\") pod \"machine-config-daemon-pndzt\" (UID: \"9695d947-fc43-4625-86af-a1a3e177bb27\") " pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.223297 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-cni-bin\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.223311 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovn-node-metrics-cert\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.223325 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovnkube-script-lib\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.244047 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.264443 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.288743 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.312368 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323725 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-ovn\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323764 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/9695d947-fc43-4625-86af-a1a3e177bb27-rootfs\") pod \"machine-config-daemon-pndzt\" (UID: \"9695d947-fc43-4625-86af-a1a3e177bb27\") " pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323778 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9695d947-fc43-4625-86af-a1a3e177bb27-proxy-tls\") pod \"machine-config-daemon-pndzt\" (UID: \"9695d947-fc43-4625-86af-a1a3e177bb27\") " pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323792 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fef52874-4a7c-4016-938a-69e88790ab19-cnibin\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323804 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-systemd-units\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323819 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jct4d\" (UniqueName: \"kubernetes.io/projected/8cd1981e-44c4-4251-81d1-c07c9d45258b-kube-api-access-jct4d\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323834 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9695d947-fc43-4625-86af-a1a3e177bb27-mcd-auth-proxy-config\") pod \"machine-config-daemon-pndzt\" (UID: \"9695d947-fc43-4625-86af-a1a3e177bb27\") " pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323847 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gg9w\" (UniqueName: \"kubernetes.io/projected/9695d947-fc43-4625-86af-a1a3e177bb27-kube-api-access-6gg9w\") pod \"machine-config-daemon-pndzt\" (UID: \"9695d947-fc43-4625-86af-a1a3e177bb27\") " pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323869 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovn-node-metrics-cert\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323883 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovnkube-script-lib\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323904 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-cni-bin\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323917 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-kubelet\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323939 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323954 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fef52874-4a7c-4016-938a-69e88790ab19-tuning-conf-dir\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323967 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-systemd\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323980 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-var-lib-openvswitch\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.323994 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-etc-openvswitch\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324007 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fef52874-4a7c-4016-938a-69e88790ab19-os-release\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324020 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-openvswitch\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324034 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-env-overrides\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324047 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fef52874-4a7c-4016-938a-69e88790ab19-system-cni-dir\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324061 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-run-netns\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324075 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-node-log\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324089 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-run-ovn-kubernetes\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324108 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-cni-netd\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324232 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fef52874-4a7c-4016-938a-69e88790ab19-cni-binary-copy\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324248 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-slash\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324263 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fef52874-4a7c-4016-938a-69e88790ab19-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324277 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-log-socket\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324300 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovnkube-config\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324315 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h29bd\" (UniqueName: \"kubernetes.io/projected/fef52874-4a7c-4016-938a-69e88790ab19-kube-api-access-h29bd\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324535 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-ovn\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.324566 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/9695d947-fc43-4625-86af-a1a3e177bb27-rootfs\") pod \"machine-config-daemon-pndzt\" (UID: \"9695d947-fc43-4625-86af-a1a3e177bb27\") " pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.325046 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-etc-openvswitch\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.325079 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fef52874-4a7c-4016-938a-69e88790ab19-cnibin\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.325111 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-cni-bin\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.325113 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-slash\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.325152 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-cni-netd\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.325237 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-log-socket\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.325724 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fef52874-4a7c-4016-938a-69e88790ab19-cni-binary-copy\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.325837 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovnkube-config\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.325878 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-systemd-units\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.325992 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fef52874-4a7c-4016-938a-69e88790ab19-os-release\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.326033 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fef52874-4a7c-4016-938a-69e88790ab19-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.326047 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovnkube-script-lib\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.326127 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-systemd\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.326139 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fef52874-4a7c-4016-938a-69e88790ab19-tuning-conf-dir\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.326156 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.326177 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-var-lib-openvswitch\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.326210 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-openvswitch\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.326214 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9695d947-fc43-4625-86af-a1a3e177bb27-mcd-auth-proxy-config\") pod \"machine-config-daemon-pndzt\" (UID: \"9695d947-fc43-4625-86af-a1a3e177bb27\") " pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.326240 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-run-netns\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.326249 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-node-log\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.326271 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fef52874-4a7c-4016-938a-69e88790ab19-system-cni-dir\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.326312 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-kubelet\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.326681 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-env-overrides\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.326889 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-run-ovn-kubernetes\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.329271 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9695d947-fc43-4625-86af-a1a3e177bb27-proxy-tls\") pod \"machine-config-daemon-pndzt\" (UID: \"9695d947-fc43-4625-86af-a1a3e177bb27\") " pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.331283 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovn-node-metrics-cert\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.331500 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.344409 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h29bd\" (UniqueName: \"kubernetes.io/projected/fef52874-4a7c-4016-938a-69e88790ab19-kube-api-access-h29bd\") pod \"multus-additional-cni-plugins-b49xq\" (UID: \"fef52874-4a7c-4016-938a-69e88790ab19\") " pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.346128 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gg9w\" (UniqueName: \"kubernetes.io/projected/9695d947-fc43-4625-86af-a1a3e177bb27-kube-api-access-6gg9w\") pod \"machine-config-daemon-pndzt\" (UID: \"9695d947-fc43-4625-86af-a1a3e177bb27\") " pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.347852 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.349001 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jct4d\" (UniqueName: \"kubernetes.io/projected/8cd1981e-44c4-4251-81d1-c07c9d45258b-kube-api-access-jct4d\") pod \"ovnkube-node-9lfpq\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.359227 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.368057 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.384408 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.395220 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.406838 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.417130 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.426582 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.437807 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.450799 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.460413 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.473454 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.482846 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.501171 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.506608 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: W1203 10:54:03.514795 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9695d947_fc43_4625_86af_a1a3e177bb27.slice/crio-e9b16003bfc1323771c0fcf5be2d658f0a9fa5aea2d28cbb2fe14a7e6c97014a WatchSource:0}: Error finding container e9b16003bfc1323771c0fcf5be2d658f0a9fa5aea2d28cbb2fe14a7e6c97014a: Status 404 returned error can't find the container with id e9b16003bfc1323771c0fcf5be2d658f0a9fa5aea2d28cbb2fe14a7e6c97014a Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.514923 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.523907 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-b49xq" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.525204 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.525317 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.525365 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.525385 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.525448 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:54:05.525421634 +0000 UTC m=+21.988477779 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.525464 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.525500 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.525515 4646 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.525520 4646 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.525500 4646 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.525554 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:05.525543068 +0000 UTC m=+21.988599203 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.525631 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:05.52562097 +0000 UTC m=+21.988677205 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.525645 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:05.5256379 +0000 UTC m=+21.988694135 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.525670 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.525787 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.525810 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.525819 4646 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.525858 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:05.525848126 +0000 UTC m=+21.988904261 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.528850 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.549712 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.573142 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.613431 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.848317 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.848430 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.848673 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.848716 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.851866 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.852782 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.854454 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.855093 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.856076 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.856620 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.857195 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.858233 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.859123 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.860227 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.860788 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.862475 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.862976 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.863566 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.864641 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.865278 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.866660 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.867218 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.867931 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.868749 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.869286 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.870504 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.871489 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.871899 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.873030 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.873507 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.874113 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.875404 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.875901 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.876965 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.877538 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.878981 4646 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.879123 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.880482 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.881136 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.881603 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.882739 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.883454 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.884819 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.885425 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.886476 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.887136 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.888034 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.888825 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.889952 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.890639 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.891606 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.892188 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.893101 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.893839 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.894725 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.895221 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.895964 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.897039 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.897642 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.898515 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.899923 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.922558 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: E1203 10:54:03.943790 4646 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfef52874_4a7c_4016_938a_69e88790ab19.slice/crio-09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500.scope\": RecentStats: unable to find data in memory cache]" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.949554 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.969751 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.971269 4646 generic.go:334] "Generic (PLEG): container finished" podID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerID="2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8" exitCode=0 Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.971356 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerDied","Data":"2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8"} Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.971420 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerStarted","Data":"ac153287e1c2e4053fbd2a176d5e16b87ddb8ab7f8fda24eca274ecd60a99269"} Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.972651 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cbp2x" event={"ID":"0b003d65-42ef-4a54-96e8-33428e114a1d","Type":"ContainerStarted","Data":"21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f"} Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.972684 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cbp2x" event={"ID":"0b003d65-42ef-4a54-96e8-33428e114a1d","Type":"ContainerStarted","Data":"bc6f438cbe1c2b95d718ffb428834bcbee7b9653178baba8fe87aeb2934c92b4"} Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.976980 4646 generic.go:334] "Generic (PLEG): container finished" podID="fef52874-4a7c-4016-938a-69e88790ab19" containerID="09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500" exitCode=0 Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.977049 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" event={"ID":"fef52874-4a7c-4016-938a-69e88790ab19","Type":"ContainerDied","Data":"09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500"} Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.977090 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" event={"ID":"fef52874-4a7c-4016-938a-69e88790ab19","Type":"ContainerStarted","Data":"88b47bb06bc707a6ed26b7c6d4ef446cd18fa5cbe5c9f1c31b54322d37cadc30"} Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.981037 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerStarted","Data":"2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb"} Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.981078 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerStarted","Data":"b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441"} Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.981115 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerStarted","Data":"e9b16003bfc1323771c0fcf5be2d658f0a9fa5aea2d28cbb2fe14a7e6c97014a"} Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.983229 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.985492 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708"} Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.986099 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:54:03 crc kubenswrapper[4646]: I1203 10:54:03.993249 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.014612 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.029379 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.057116 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.081295 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.097666 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.123936 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.140760 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.171285 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.184474 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.221225 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.233822 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.245057 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.257575 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.272914 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.289576 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.306367 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.321217 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.335558 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.359066 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.390054 4646 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.391504 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.391531 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.391540 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.391623 4646 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.404788 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.451411 4646 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.451647 4646 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.453223 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.453246 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.453253 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.453266 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.453275 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:04Z","lastTransitionTime":"2025-12-03T10:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:04 crc kubenswrapper[4646]: E1203 10:54:04.470482 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.473513 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.473536 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.473545 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.473558 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.473566 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:04Z","lastTransitionTime":"2025-12-03T10:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.481906 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: E1203 10:54:04.488196 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.492506 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.492535 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.492547 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.492563 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.492572 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:04Z","lastTransitionTime":"2025-12-03T10:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:04 crc kubenswrapper[4646]: E1203 10:54:04.503985 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.506721 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.506749 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.506758 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.506771 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.506779 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:04Z","lastTransitionTime":"2025-12-03T10:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:04 crc kubenswrapper[4646]: E1203 10:54:04.518204 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.521090 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.521111 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.521118 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.521133 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.521141 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:04Z","lastTransitionTime":"2025-12-03T10:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:04 crc kubenswrapper[4646]: E1203 10:54:04.548976 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:04 crc kubenswrapper[4646]: E1203 10:54:04.549087 4646 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.550328 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.550375 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.550383 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.550395 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.550404 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:04Z","lastTransitionTime":"2025-12-03T10:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.652807 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.652839 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.652849 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.652864 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.652875 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:04Z","lastTransitionTime":"2025-12-03T10:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.755496 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.755530 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.755543 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.755560 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.755572 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:04Z","lastTransitionTime":"2025-12-03T10:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.847418 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:04 crc kubenswrapper[4646]: E1203 10:54:04.847534 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.860654 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.860924 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.860934 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.860949 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.860958 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:04Z","lastTransitionTime":"2025-12-03T10:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.963476 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.963516 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.963524 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.963576 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.963585 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:04Z","lastTransitionTime":"2025-12-03T10:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.990260 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18"} Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.993095 4646 generic.go:334] "Generic (PLEG): container finished" podID="fef52874-4a7c-4016-938a-69e88790ab19" containerID="5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5" exitCode=0 Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.993161 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" event={"ID":"fef52874-4a7c-4016-938a-69e88790ab19","Type":"ContainerDied","Data":"5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5"} Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.998452 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerStarted","Data":"b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8"} Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.998508 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerStarted","Data":"a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca"} Dec 03 10:54:04 crc kubenswrapper[4646]: I1203 10:54:04.998522 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerStarted","Data":"71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88"} Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.012786 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.024174 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.045265 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.061894 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.065568 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.065727 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.065740 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.065758 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.065770 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:05Z","lastTransitionTime":"2025-12-03T10:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.072919 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.083622 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.097306 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.115158 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.136454 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.150370 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.163447 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.171762 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.171800 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.171811 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.171827 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.171842 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:05Z","lastTransitionTime":"2025-12-03T10:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.191042 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.206074 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.215638 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.229994 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.242379 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.255253 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.270183 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.274111 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.274208 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.274263 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.274319 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.274386 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:05Z","lastTransitionTime":"2025-12-03T10:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.286890 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.307031 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.319298 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.359778 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.376440 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.376470 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.376479 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.376494 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.376503 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:05Z","lastTransitionTime":"2025-12-03T10:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.399934 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.443850 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.478950 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.479224 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.479244 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.479252 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.479265 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.479274 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:05Z","lastTransitionTime":"2025-12-03T10:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.523515 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.546496 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.546649 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:05 crc kubenswrapper[4646]: E1203 10:54:05.546702 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:54:09.546669417 +0000 UTC m=+26.009725572 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.546759 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.546807 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:05 crc kubenswrapper[4646]: E1203 10:54:05.546763 4646 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.546870 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:05 crc kubenswrapper[4646]: E1203 10:54:05.546947 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:54:05 crc kubenswrapper[4646]: E1203 10:54:05.546962 4646 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:54:05 crc kubenswrapper[4646]: E1203 10:54:05.546974 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:54:05 crc kubenswrapper[4646]: E1203 10:54:05.547008 4646 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:05 crc kubenswrapper[4646]: E1203 10:54:05.546870 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:54:05 crc kubenswrapper[4646]: E1203 10:54:05.547071 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:54:05 crc kubenswrapper[4646]: E1203 10:54:05.547084 4646 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:05 crc kubenswrapper[4646]: E1203 10:54:05.547006 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:09.546966766 +0000 UTC m=+26.010022941 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:54:05 crc kubenswrapper[4646]: E1203 10:54:05.547180 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:09.547139681 +0000 UTC m=+26.010195856 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:54:05 crc kubenswrapper[4646]: E1203 10:54:05.547219 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:09.547200523 +0000 UTC m=+26.010256698 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:05 crc kubenswrapper[4646]: E1203 10:54:05.547246 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:09.547233093 +0000 UTC m=+26.010289268 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.562542 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.582937 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.582978 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.582989 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.583007 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.583019 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:05Z","lastTransitionTime":"2025-12-03T10:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.606038 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:05Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.686208 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.686252 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.686263 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.686281 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.686293 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:05Z","lastTransitionTime":"2025-12-03T10:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.789047 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.789095 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.789115 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.789139 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.789157 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:05Z","lastTransitionTime":"2025-12-03T10:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.847502 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:05 crc kubenswrapper[4646]: E1203 10:54:05.847702 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.847850 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:05 crc kubenswrapper[4646]: E1203 10:54:05.847971 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.892633 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.892707 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.892724 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.892749 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.892766 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:05Z","lastTransitionTime":"2025-12-03T10:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.996207 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.996265 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.996278 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.996299 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:05 crc kubenswrapper[4646]: I1203 10:54:05.996316 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:05Z","lastTransitionTime":"2025-12-03T10:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.005772 4646 generic.go:334] "Generic (PLEG): container finished" podID="fef52874-4a7c-4016-938a-69e88790ab19" containerID="5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb" exitCode=0 Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.005881 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" event={"ID":"fef52874-4a7c-4016-938a-69e88790ab19","Type":"ContainerDied","Data":"5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb"} Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.011513 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerStarted","Data":"7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104"} Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.011574 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerStarted","Data":"9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616"} Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.011596 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerStarted","Data":"b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119"} Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.029628 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.048688 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.064263 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.080514 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.093274 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.101904 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.101933 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.101942 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.101956 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.101965 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:06Z","lastTransitionTime":"2025-12-03T10:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.119904 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.140734 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.153811 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.167913 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.181159 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.194756 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.203850 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.203893 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.203905 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.203923 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.203934 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:06Z","lastTransitionTime":"2025-12-03T10:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.205549 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.217410 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.225781 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.306494 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.306538 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.306555 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.306572 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.306584 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:06Z","lastTransitionTime":"2025-12-03T10:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.408646 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.408690 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.408703 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.408720 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.408730 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:06Z","lastTransitionTime":"2025-12-03T10:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.511585 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.511644 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.511660 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.511691 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.511708 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:06Z","lastTransitionTime":"2025-12-03T10:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.613813 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.613865 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.613882 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.613906 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.613920 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:06Z","lastTransitionTime":"2025-12-03T10:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.717185 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.717239 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.717257 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.717296 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.717316 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:06Z","lastTransitionTime":"2025-12-03T10:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.821128 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.821164 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.821173 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.821186 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.821195 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:06Z","lastTransitionTime":"2025-12-03T10:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.847968 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:06 crc kubenswrapper[4646]: E1203 10:54:06.848085 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.923474 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.923555 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.923581 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.923706 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:06 crc kubenswrapper[4646]: I1203 10:54:06.923781 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:06Z","lastTransitionTime":"2025-12-03T10:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.018744 4646 generic.go:334] "Generic (PLEG): container finished" podID="fef52874-4a7c-4016-938a-69e88790ab19" containerID="268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e" exitCode=0 Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.018795 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" event={"ID":"fef52874-4a7c-4016-938a-69e88790ab19","Type":"ContainerDied","Data":"268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e"} Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.027100 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.027147 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.027162 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.027182 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.027197 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:07Z","lastTransitionTime":"2025-12-03T10:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.038545 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:07Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.057497 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:07Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.074404 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:07Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.086090 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:07Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.101285 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:07Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.112264 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:07Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.127367 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:07Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.129305 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.129354 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.129364 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.129378 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.129394 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:07Z","lastTransitionTime":"2025-12-03T10:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.143713 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:07Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.160061 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:07Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.175143 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:07Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.189956 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:07Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.201429 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:07Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.219634 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:07Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.231605 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:07Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.232306 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.232327 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.232357 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.232372 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.232381 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:07Z","lastTransitionTime":"2025-12-03T10:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.335396 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.335468 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.335495 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.335525 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.335545 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:07Z","lastTransitionTime":"2025-12-03T10:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.438933 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.439006 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.439033 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.439066 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.439091 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:07Z","lastTransitionTime":"2025-12-03T10:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.541790 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.541854 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.541865 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.541904 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.541920 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:07Z","lastTransitionTime":"2025-12-03T10:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.645281 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.645325 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.645356 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.645374 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.645385 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:07Z","lastTransitionTime":"2025-12-03T10:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.748123 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.748178 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.748198 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.748230 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.748250 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:07Z","lastTransitionTime":"2025-12-03T10:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.847826 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:07 crc kubenswrapper[4646]: E1203 10:54:07.847992 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.848375 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:07 crc kubenswrapper[4646]: E1203 10:54:07.848457 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.850244 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.850280 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.850291 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.850308 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.850317 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:07Z","lastTransitionTime":"2025-12-03T10:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.952534 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.952588 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.952625 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.952653 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:07 crc kubenswrapper[4646]: I1203 10:54:07.952671 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:07Z","lastTransitionTime":"2025-12-03T10:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.028299 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerStarted","Data":"ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe"} Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.032752 4646 generic.go:334] "Generic (PLEG): container finished" podID="fef52874-4a7c-4016-938a-69e88790ab19" containerID="9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817" exitCode=0 Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.032806 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" event={"ID":"fef52874-4a7c-4016-938a-69e88790ab19","Type":"ContainerDied","Data":"9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817"} Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.052440 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:08Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.056633 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.056693 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.056717 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.056748 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.056772 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:08Z","lastTransitionTime":"2025-12-03T10:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.081244 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:08Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.100042 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:08Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.114237 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:08Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.127414 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:08Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.149844 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:08Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.159394 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.159446 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.159461 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.159481 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.159502 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:08Z","lastTransitionTime":"2025-12-03T10:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.171075 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:08Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.187130 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:08Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.198075 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:08Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.211773 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:08Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.226656 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:08Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.242874 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:08Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.253070 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:08Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.262501 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.262538 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.262559 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.262576 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.262588 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:08Z","lastTransitionTime":"2025-12-03T10:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.269059 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:08Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.364720 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.364767 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.364784 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.364804 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.364818 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:08Z","lastTransitionTime":"2025-12-03T10:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.468132 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.468198 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.468216 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.468244 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.468279 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:08Z","lastTransitionTime":"2025-12-03T10:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.570964 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.571017 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.571044 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.571068 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.571084 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:08Z","lastTransitionTime":"2025-12-03T10:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.672724 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.672757 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.672764 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.672776 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.672786 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:08Z","lastTransitionTime":"2025-12-03T10:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.775708 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.775800 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.775840 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.775877 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.775901 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:08Z","lastTransitionTime":"2025-12-03T10:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.847911 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:08 crc kubenswrapper[4646]: E1203 10:54:08.848052 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.878867 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.878902 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.878914 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.878937 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.878960 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:08Z","lastTransitionTime":"2025-12-03T10:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.982099 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.982152 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.982169 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.982191 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:08 crc kubenswrapper[4646]: I1203 10:54:08.982206 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:08Z","lastTransitionTime":"2025-12-03T10:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.039212 4646 generic.go:334] "Generic (PLEG): container finished" podID="fef52874-4a7c-4016-938a-69e88790ab19" containerID="33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9" exitCode=0 Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.039254 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" event={"ID":"fef52874-4a7c-4016-938a-69e88790ab19","Type":"ContainerDied","Data":"33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9"} Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.055178 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:09Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.069203 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:09Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.081962 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:09Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.086529 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.086562 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.086574 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.086590 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.086602 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:09Z","lastTransitionTime":"2025-12-03T10:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.098767 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:09Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.112260 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:09Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.122464 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:09Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.135461 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:09Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.149274 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:09Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.168593 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:09Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.183486 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:09Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.189270 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.189302 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.189311 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.189325 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.189347 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:09Z","lastTransitionTime":"2025-12-03T10:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.197141 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:09Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.207677 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:09Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.218837 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:09Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.248981 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:09Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.291950 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.291995 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.292008 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.292026 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.292039 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:09Z","lastTransitionTime":"2025-12-03T10:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.394190 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.394229 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.394239 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.394255 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.394267 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:09Z","lastTransitionTime":"2025-12-03T10:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.496585 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.496787 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.496872 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.496938 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.497012 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:09Z","lastTransitionTime":"2025-12-03T10:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.586517 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.586658 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:09 crc kubenswrapper[4646]: E1203 10:54:09.586767 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:54:17.586705313 +0000 UTC m=+34.049761488 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:54:09 crc kubenswrapper[4646]: E1203 10:54:09.586779 4646 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.586830 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:09 crc kubenswrapper[4646]: E1203 10:54:09.586846 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:17.586827707 +0000 UTC m=+34.049883872 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.586896 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.586989 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:09 crc kubenswrapper[4646]: E1203 10:54:09.587016 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:54:09 crc kubenswrapper[4646]: E1203 10:54:09.587053 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:54:09 crc kubenswrapper[4646]: E1203 10:54:09.587070 4646 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:09 crc kubenswrapper[4646]: E1203 10:54:09.587139 4646 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:54:09 crc kubenswrapper[4646]: E1203 10:54:09.587151 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:17.587117705 +0000 UTC m=+34.050173910 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:09 crc kubenswrapper[4646]: E1203 10:54:09.587225 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:54:09 crc kubenswrapper[4646]: E1203 10:54:09.587234 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:17.587211798 +0000 UTC m=+34.050267993 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:54:09 crc kubenswrapper[4646]: E1203 10:54:09.587244 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:54:09 crc kubenswrapper[4646]: E1203 10:54:09.587257 4646 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:09 crc kubenswrapper[4646]: E1203 10:54:09.587296 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:17.5872836 +0000 UTC m=+34.050339835 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.599482 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.599527 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.599539 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.599574 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.599602 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:09Z","lastTransitionTime":"2025-12-03T10:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.709878 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.709920 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.709938 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.709966 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.709984 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:09Z","lastTransitionTime":"2025-12-03T10:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.811987 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.812035 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.812055 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.812080 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.812099 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:09Z","lastTransitionTime":"2025-12-03T10:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.847808 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.847838 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:09 crc kubenswrapper[4646]: E1203 10:54:09.847923 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:09 crc kubenswrapper[4646]: E1203 10:54:09.848141 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.915593 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.915663 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.915681 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.915706 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:09 crc kubenswrapper[4646]: I1203 10:54:09.915725 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:09Z","lastTransitionTime":"2025-12-03T10:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.018570 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.018928 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.018939 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.018958 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.018969 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:10Z","lastTransitionTime":"2025-12-03T10:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.049029 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" event={"ID":"fef52874-4a7c-4016-938a-69e88790ab19","Type":"ContainerStarted","Data":"39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504"} Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.057626 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerStarted","Data":"cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8"} Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.057962 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.072317 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.085443 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.098913 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.099966 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.112731 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.121497 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.121541 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.121553 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.121570 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.121584 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:10Z","lastTransitionTime":"2025-12-03T10:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.132379 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.151294 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.164298 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.175163 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.184768 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.195002 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.211150 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.222761 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.224218 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.224274 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.224375 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.224422 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.224494 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:10Z","lastTransitionTime":"2025-12-03T10:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.238914 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.249245 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.262241 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.273988 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.298236 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.319004 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.327383 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.327449 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.327487 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.327528 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.327554 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:10Z","lastTransitionTime":"2025-12-03T10:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.334159 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.350224 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.365485 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.381917 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.398246 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.410816 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.426032 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.429673 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.429719 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.429735 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.429757 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.429772 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:10Z","lastTransitionTime":"2025-12-03T10:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.439441 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.452430 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.462525 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:10Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.531488 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.531529 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.531541 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.531560 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.531573 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:10Z","lastTransitionTime":"2025-12-03T10:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.633662 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.633710 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.633719 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.633753 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.633763 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:10Z","lastTransitionTime":"2025-12-03T10:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.737109 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.737176 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.737203 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.737233 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.737258 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:10Z","lastTransitionTime":"2025-12-03T10:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.840209 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.840272 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.840290 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.840318 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.840377 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:10Z","lastTransitionTime":"2025-12-03T10:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.847771 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:10 crc kubenswrapper[4646]: E1203 10:54:10.847949 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.943221 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.943268 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.943278 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.943294 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:10 crc kubenswrapper[4646]: I1203 10:54:10.943308 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:10Z","lastTransitionTime":"2025-12-03T10:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.046985 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.047031 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.047048 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.047067 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.047081 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:11Z","lastTransitionTime":"2025-12-03T10:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.061522 4646 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.062106 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.101641 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.122880 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:11Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.140655 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:11Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.149425 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.149490 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.149502 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.149521 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.149545 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:11Z","lastTransitionTime":"2025-12-03T10:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.165530 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:11Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.181284 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:11Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.207953 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:11Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.237731 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:11Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.252403 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.252439 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.252447 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.252460 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.252470 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:11Z","lastTransitionTime":"2025-12-03T10:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.266602 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:11Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.286451 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:11Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.296608 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:11Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.307872 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:11Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.317616 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:11Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.333711 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:11Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.345988 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:11Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.354729 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.354761 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.354776 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.354799 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.354813 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:11Z","lastTransitionTime":"2025-12-03T10:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.360924 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:11Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.457531 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.457578 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.457596 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.457625 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.457649 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:11Z","lastTransitionTime":"2025-12-03T10:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.560387 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.560415 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.560423 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.560445 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.560455 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:11Z","lastTransitionTime":"2025-12-03T10:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.662666 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.662704 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.662713 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.662732 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.662743 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:11Z","lastTransitionTime":"2025-12-03T10:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.764846 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.764892 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.764906 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.764923 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.764934 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:11Z","lastTransitionTime":"2025-12-03T10:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.847652 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:11 crc kubenswrapper[4646]: E1203 10:54:11.847864 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.848400 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:11 crc kubenswrapper[4646]: E1203 10:54:11.848536 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.893261 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.893287 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.893295 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.893308 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.893316 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:11Z","lastTransitionTime":"2025-12-03T10:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.996494 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.996531 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.996538 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.996554 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:11 crc kubenswrapper[4646]: I1203 10:54:11.996564 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:11Z","lastTransitionTime":"2025-12-03T10:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.064559 4646 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.099381 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.099416 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.099427 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.099440 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.099448 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:12Z","lastTransitionTime":"2025-12-03T10:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.202009 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.202053 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.202064 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.202082 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.202094 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:12Z","lastTransitionTime":"2025-12-03T10:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.305131 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.305168 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.305181 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.305198 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.305211 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:12Z","lastTransitionTime":"2025-12-03T10:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.409893 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.409944 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.409958 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.409982 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.410003 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:12Z","lastTransitionTime":"2025-12-03T10:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.512547 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.512640 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.512665 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.512694 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.512717 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:12Z","lastTransitionTime":"2025-12-03T10:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.616612 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.616648 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.616660 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.616677 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.616692 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:12Z","lastTransitionTime":"2025-12-03T10:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.719533 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.719593 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.719611 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.719636 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.719654 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:12Z","lastTransitionTime":"2025-12-03T10:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.822585 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.822636 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.822653 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.822676 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.822692 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:12Z","lastTransitionTime":"2025-12-03T10:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.847892 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:12 crc kubenswrapper[4646]: E1203 10:54:12.848065 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.925813 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.925865 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.925883 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.925908 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:12 crc kubenswrapper[4646]: I1203 10:54:12.925925 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:12Z","lastTransitionTime":"2025-12-03T10:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.028960 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.029014 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.029039 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.029068 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.029089 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:13Z","lastTransitionTime":"2025-12-03T10:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.070889 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovnkube-controller/0.log" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.075696 4646 generic.go:334] "Generic (PLEG): container finished" podID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerID="cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8" exitCode=1 Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.075764 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerDied","Data":"cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8"} Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.077236 4646 scope.go:117] "RemoveContainer" containerID="cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.099835 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.119666 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.131189 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.131233 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.131257 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.131274 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.131285 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:13Z","lastTransitionTime":"2025-12-03T10:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.137595 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.157881 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.173691 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.193863 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:12Z\\\",\\\"message\\\":\\\"203 10:54:12.322855 5849 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 10:54:12.323213 5849 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 10:54:12.323212 5849 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:12.323244 5849 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:12.323254 5849 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:12.323265 5849 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:12.323282 5849 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:12.323285 5849 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 10:54:12.323302 5849 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 10:54:12.323306 5849 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:12.323309 5849 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:12.323313 5849 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 10:54:12.323324 5849 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:12.323352 5849 factory.go:656] Stopping watch factory\\\\nI1203 10:54:12.323354 5849 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:12.323363 5849 ovnkube.go:599] Stopped ovnkube\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.214194 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.230322 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.233785 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.233824 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.233833 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.233848 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.233860 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:13Z","lastTransitionTime":"2025-12-03T10:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.241666 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.253860 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.269848 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.278893 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.289045 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.299770 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.329714 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.335846 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.335887 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.335898 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.335915 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.335928 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:13Z","lastTransitionTime":"2025-12-03T10:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.346219 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.360862 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.375589 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.388851 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.400489 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.418653 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.435060 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.438959 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.438996 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.439005 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.439019 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.439027 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:13Z","lastTransitionTime":"2025-12-03T10:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.446849 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.456500 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.474328 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:12Z\\\",\\\"message\\\":\\\"203 10:54:12.322855 5849 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 10:54:12.323213 5849 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 10:54:12.323212 5849 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:12.323244 5849 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:12.323254 5849 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:12.323265 5849 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:12.323282 5849 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:12.323285 5849 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 10:54:12.323302 5849 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 10:54:12.323306 5849 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:12.323309 5849 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:12.323313 5849 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 10:54:12.323324 5849 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:12.323352 5849 factory.go:656] Stopping watch factory\\\\nI1203 10:54:12.323354 5849 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:12.323363 5849 ovnkube.go:599] Stopped ovnkube\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.492384 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.505544 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.523356 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.536710 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.540973 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.541026 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.541040 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.541059 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.541072 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:13Z","lastTransitionTime":"2025-12-03T10:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.642803 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.643049 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.643160 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.643271 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.643540 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:13Z","lastTransitionTime":"2025-12-03T10:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.746680 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.746720 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.746731 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.746746 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.746756 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:13Z","lastTransitionTime":"2025-12-03T10:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.847639 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:13 crc kubenswrapper[4646]: E1203 10:54:13.847741 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.848043 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:13 crc kubenswrapper[4646]: E1203 10:54:13.848097 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.849562 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.849628 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.849644 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.849675 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.849694 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:13Z","lastTransitionTime":"2025-12-03T10:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.880590 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.896410 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.919309 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.936458 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.947475 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.952168 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.952220 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.952228 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.952245 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.952255 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:13Z","lastTransitionTime":"2025-12-03T10:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.972652 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:12Z\\\",\\\"message\\\":\\\"203 10:54:12.322855 5849 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 10:54:12.323213 5849 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 10:54:12.323212 5849 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:12.323244 5849 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:12.323254 5849 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:12.323265 5849 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:12.323282 5849 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:12.323285 5849 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 10:54:12.323302 5849 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 10:54:12.323306 5849 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:12.323309 5849 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:12.323313 5849 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 10:54:12.323324 5849 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:12.323352 5849 factory.go:656] Stopping watch factory\\\\nI1203 10:54:12.323354 5849 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:12.323363 5849 ovnkube.go:599] Stopped ovnkube\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: I1203 10:54:13.990165 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:13 crc kubenswrapper[4646]: E1203 10:54:13.995851 4646 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cd1981e_44c4_4251_81d1_c07c9d45258b.slice/crio-4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff.scope\": RecentStats: unable to find data in memory cache]" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.000762 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:13Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.010502 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.021657 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.039401 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.054307 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.054369 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.054383 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.054402 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.054657 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:14Z","lastTransitionTime":"2025-12-03T10:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.080359 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovnkube-controller/0.log" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.085238 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerStarted","Data":"4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff"} Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.085438 4646 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.107539 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.121061 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.122306 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.130454 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.143989 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.157072 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.160922 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.160964 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.160978 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.160999 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.161013 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:14Z","lastTransitionTime":"2025-12-03T10:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.176549 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.187763 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.198911 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.214157 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.225437 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.244618 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.262712 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.263655 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.263695 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.263708 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.263727 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.263739 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:14Z","lastTransitionTime":"2025-12-03T10:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.275887 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.285715 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.296104 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.305627 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.320835 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:12Z\\\",\\\"message\\\":\\\"203 10:54:12.322855 5849 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 10:54:12.323213 5849 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 10:54:12.323212 5849 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:12.323244 5849 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:12.323254 5849 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:12.323265 5849 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:12.323282 5849 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:12.323285 5849 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 10:54:12.323302 5849 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 10:54:12.323306 5849 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:12.323309 5849 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:12.323313 5849 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 10:54:12.323324 5849 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:12.323352 5849 factory.go:656] Stopping watch factory\\\\nI1203 10:54:12.323354 5849 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:12.323363 5849 ovnkube.go:599] Stopped ovnkube\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.365918 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.366209 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.366288 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.366381 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.366441 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:14Z","lastTransitionTime":"2025-12-03T10:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.469945 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.469991 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.470005 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.470023 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.470035 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:14Z","lastTransitionTime":"2025-12-03T10:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.572150 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.572184 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.572194 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.572207 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.572215 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:14Z","lastTransitionTime":"2025-12-03T10:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.674394 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.674424 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.674432 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.674445 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.674454 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:14Z","lastTransitionTime":"2025-12-03T10:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.777538 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.777575 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.777587 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.777606 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.777619 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:14Z","lastTransitionTime":"2025-12-03T10:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.847738 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:14 crc kubenswrapper[4646]: E1203 10:54:14.847917 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.880528 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.880588 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.880610 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.880635 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.880655 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:14Z","lastTransitionTime":"2025-12-03T10:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.883241 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.883266 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.883273 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.883287 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.883295 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:14Z","lastTransitionTime":"2025-12-03T10:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:14 crc kubenswrapper[4646]: E1203 10:54:14.896235 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.900773 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.900837 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.900855 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.900882 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.900899 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:14Z","lastTransitionTime":"2025-12-03T10:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:14 crc kubenswrapper[4646]: E1203 10:54:14.916771 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.921414 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.921464 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.921475 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.921492 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.921855 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:14Z","lastTransitionTime":"2025-12-03T10:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:14 crc kubenswrapper[4646]: E1203 10:54:14.941136 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.945405 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.945509 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.945524 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.945542 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.945554 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:14Z","lastTransitionTime":"2025-12-03T10:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:14 crc kubenswrapper[4646]: E1203 10:54:14.961546 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.968817 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.968963 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.968981 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.969005 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.969028 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:14Z","lastTransitionTime":"2025-12-03T10:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:14 crc kubenswrapper[4646]: E1203 10:54:14.987371 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:14Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:14 crc kubenswrapper[4646]: E1203 10:54:14.987839 4646 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.989803 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.990260 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.990280 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.990304 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.990319 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:14Z","lastTransitionTime":"2025-12-03T10:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.995606 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8"] Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.996195 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" Dec 03 10:54:14 crc kubenswrapper[4646]: I1203 10:54:14.998087 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.001607 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.011475 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.025726 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.041232 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.041676 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/263a8752-97c7-4d08-9822-1e85687116fd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-gprb8\" (UID: \"263a8752-97c7-4d08-9822-1e85687116fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.041732 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rhdl\" (UniqueName: \"kubernetes.io/projected/263a8752-97c7-4d08-9822-1e85687116fd-kube-api-access-5rhdl\") pod \"ovnkube-control-plane-749d76644c-gprb8\" (UID: \"263a8752-97c7-4d08-9822-1e85687116fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.041783 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/263a8752-97c7-4d08-9822-1e85687116fd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-gprb8\" (UID: \"263a8752-97c7-4d08-9822-1e85687116fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.041818 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/263a8752-97c7-4d08-9822-1e85687116fd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-gprb8\" (UID: \"263a8752-97c7-4d08-9822-1e85687116fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.057889 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.069651 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.082669 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.090117 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovnkube-controller/1.log" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.090741 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovnkube-controller/0.log" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.093169 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.093269 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.093287 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.093351 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.093429 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:15Z","lastTransitionTime":"2025-12-03T10:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.094221 4646 generic.go:334] "Generic (PLEG): container finished" podID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerID="4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff" exitCode=1 Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.094290 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerDied","Data":"4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff"} Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.094477 4646 scope.go:117] "RemoveContainer" containerID="cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.095780 4646 scope.go:117] "RemoveContainer" containerID="4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff" Dec 03 10:54:15 crc kubenswrapper[4646]: E1203 10:54:15.096116 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.101239 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.119247 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.134768 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.142790 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/263a8752-97c7-4d08-9822-1e85687116fd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-gprb8\" (UID: \"263a8752-97c7-4d08-9822-1e85687116fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.143437 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rhdl\" (UniqueName: \"kubernetes.io/projected/263a8752-97c7-4d08-9822-1e85687116fd-kube-api-access-5rhdl\") pod \"ovnkube-control-plane-749d76644c-gprb8\" (UID: \"263a8752-97c7-4d08-9822-1e85687116fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.143534 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/263a8752-97c7-4d08-9822-1e85687116fd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-gprb8\" (UID: \"263a8752-97c7-4d08-9822-1e85687116fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.143630 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/263a8752-97c7-4d08-9822-1e85687116fd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-gprb8\" (UID: \"263a8752-97c7-4d08-9822-1e85687116fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.144671 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/263a8752-97c7-4d08-9822-1e85687116fd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-gprb8\" (UID: \"263a8752-97c7-4d08-9822-1e85687116fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.144925 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/263a8752-97c7-4d08-9822-1e85687116fd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-gprb8\" (UID: \"263a8752-97c7-4d08-9822-1e85687116fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.151492 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/263a8752-97c7-4d08-9822-1e85687116fd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-gprb8\" (UID: \"263a8752-97c7-4d08-9822-1e85687116fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.151913 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.161983 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rhdl\" (UniqueName: \"kubernetes.io/projected/263a8752-97c7-4d08-9822-1e85687116fd-kube-api-access-5rhdl\") pod \"ovnkube-control-plane-749d76644c-gprb8\" (UID: \"263a8752-97c7-4d08-9822-1e85687116fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.185387 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.195910 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.196174 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.196260 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.196373 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.196472 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:15Z","lastTransitionTime":"2025-12-03T10:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.222460 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.253708 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:12Z\\\",\\\"message\\\":\\\"203 10:54:12.322855 5849 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 10:54:12.323213 5849 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 10:54:12.323212 5849 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:12.323244 5849 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:12.323254 5849 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:12.323265 5849 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:12.323282 5849 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:12.323285 5849 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 10:54:12.323302 5849 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 10:54:12.323306 5849 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:12.323309 5849 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:12.323313 5849 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 10:54:12.323324 5849 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:12.323352 5849 factory.go:656] Stopping watch factory\\\\nI1203 10:54:12.323354 5849 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:12.323363 5849 ovnkube.go:599] Stopped ovnkube\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.268373 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.282608 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.293549 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.299104 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.299148 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.299162 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.299201 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.299215 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:15Z","lastTransitionTime":"2025-12-03T10:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.303666 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.312545 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.317180 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.336049 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.353452 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.366531 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.390104 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.401751 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.402318 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.402363 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.402373 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.402389 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.402399 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:15Z","lastTransitionTime":"2025-12-03T10:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.416473 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.434146 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.449403 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.464505 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.478531 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.490850 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.504485 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.504513 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.504523 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.504539 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.504550 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:15Z","lastTransitionTime":"2025-12-03T10:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.507893 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc9c6459dcd5b0a6c3d0cc89942c58ddfbd0b9c2c13551e4b3e9a10e39c4d2b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:12Z\\\",\\\"message\\\":\\\"203 10:54:12.322855 5849 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 10:54:12.323213 5849 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 10:54:12.323212 5849 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:12.323244 5849 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:12.323254 5849 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:12.323265 5849 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:12.323282 5849 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:12.323285 5849 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 10:54:12.323302 5849 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 10:54:12.323306 5849 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:12.323309 5849 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:12.323313 5849 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 10:54:12.323324 5849 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:12.323352 5849 factory.go:656] Stopping watch factory\\\\nI1203 10:54:12.323354 5849 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:12.323363 5849 ovnkube.go:599] Stopped ovnkube\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:13.933771 5966 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.933938 5966 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934286 5966 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934436 5966 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934859 5966 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.935138 5966 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.968608 5966 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1203 10:54:13.968680 5966 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1203 10:54:13.968829 5966 ovnkube.go:599] Stopped ovnkube\\\\nI1203 10:54:13.968859 5966 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 10:54:13.968971 5966 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:15Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.607953 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.608003 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.608015 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.608038 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.608049 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:15Z","lastTransitionTime":"2025-12-03T10:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.710840 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.710886 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.710896 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.710913 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.710926 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:15Z","lastTransitionTime":"2025-12-03T10:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.813857 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.813927 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.813938 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.813961 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.813976 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:15Z","lastTransitionTime":"2025-12-03T10:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.848163 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.848166 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:15 crc kubenswrapper[4646]: E1203 10:54:15.848389 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:15 crc kubenswrapper[4646]: E1203 10:54:15.848505 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.918549 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.918588 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.918599 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.918617 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:15 crc kubenswrapper[4646]: I1203 10:54:15.918630 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:15Z","lastTransitionTime":"2025-12-03T10:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.021208 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.021256 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.021268 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.021285 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.021297 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:16Z","lastTransitionTime":"2025-12-03T10:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.099712 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovnkube-controller/1.log" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.104576 4646 scope.go:117] "RemoveContainer" containerID="4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff" Dec 03 10:54:16 crc kubenswrapper[4646]: E1203 10:54:16.104833 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.106071 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" event={"ID":"263a8752-97c7-4d08-9822-1e85687116fd","Type":"ContainerStarted","Data":"143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d"} Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.106132 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" event={"ID":"263a8752-97c7-4d08-9822-1e85687116fd","Type":"ContainerStarted","Data":"55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671"} Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.106154 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" event={"ID":"263a8752-97c7-4d08-9822-1e85687116fd","Type":"ContainerStarted","Data":"47a8669d88243054394a5acf397a4c6664b994cf904cff1979ddd96aa876445c"} Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.121291 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.123580 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.123617 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.123628 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.123650 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.123662 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:16Z","lastTransitionTime":"2025-12-03T10:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.136786 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.149664 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.166151 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.181482 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.203174 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.214732 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.226752 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.226795 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.226809 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.226830 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.226845 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:16Z","lastTransitionTime":"2025-12-03T10:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.228037 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.241019 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.260510 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:13.933771 5966 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.933938 5966 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934286 5966 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934436 5966 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934859 5966 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.935138 5966 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.968608 5966 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1203 10:54:13.968680 5966 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1203 10:54:13.968829 5966 ovnkube.go:599] Stopped ovnkube\\\\nI1203 10:54:13.968859 5966 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 10:54:13.968971 5966 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.274867 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.287220 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.303130 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.321418 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.329977 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.330022 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.330039 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.330062 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.330079 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:16Z","lastTransitionTime":"2025-12-03T10:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.334935 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.348243 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.359322 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.373097 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.389792 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.407787 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.418894 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.433282 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.433359 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.433373 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.433396 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.433408 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:16Z","lastTransitionTime":"2025-12-03T10:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.434292 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.448376 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.470233 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.477655 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-rs5rq"] Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.478286 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:16 crc kubenswrapper[4646]: E1203 10:54:16.478402 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.485622 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.501149 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.515615 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.532112 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.535650 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.535679 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.535687 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.535702 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.535711 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:16Z","lastTransitionTime":"2025-12-03T10:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.552081 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:13.933771 5966 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.933938 5966 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934286 5966 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934436 5966 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934859 5966 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.935138 5966 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.968608 5966 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1203 10:54:13.968680 5966 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1203 10:54:13.968829 5966 ovnkube.go:599] Stopped ovnkube\\\\nI1203 10:54:13.968859 5966 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 10:54:13.968971 5966 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.558813 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs\") pod \"network-metrics-daemon-rs5rq\" (UID: \"51b171e4-cc29-4118-908c-d13825362e1c\") " pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.558892 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8kkt\" (UniqueName: \"kubernetes.io/projected/51b171e4-cc29-4118-908c-d13825362e1c-kube-api-access-m8kkt\") pod \"network-metrics-daemon-rs5rq\" (UID: \"51b171e4-cc29-4118-908c-d13825362e1c\") " pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.569786 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.583792 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.596729 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.609237 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b171e4-cc29-4118-908c-d13825362e1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.622614 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.634479 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.637785 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.637848 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.637858 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.637872 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.637882 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:16Z","lastTransitionTime":"2025-12-03T10:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.650644 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.660071 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs\") pod \"network-metrics-daemon-rs5rq\" (UID: \"51b171e4-cc29-4118-908c-d13825362e1c\") " pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.660107 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8kkt\" (UniqueName: \"kubernetes.io/projected/51b171e4-cc29-4118-908c-d13825362e1c-kube-api-access-m8kkt\") pod \"network-metrics-daemon-rs5rq\" (UID: \"51b171e4-cc29-4118-908c-d13825362e1c\") " pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:16 crc kubenswrapper[4646]: E1203 10:54:16.660212 4646 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:54:16 crc kubenswrapper[4646]: E1203 10:54:16.660273 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs podName:51b171e4-cc29-4118-908c-d13825362e1c nodeName:}" failed. No retries permitted until 2025-12-03 10:54:17.160256009 +0000 UTC m=+33.623312144 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs") pod "network-metrics-daemon-rs5rq" (UID: "51b171e4-cc29-4118-908c-d13825362e1c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.664162 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.676604 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8kkt\" (UniqueName: \"kubernetes.io/projected/51b171e4-cc29-4118-908c-d13825362e1c-kube-api-access-m8kkt\") pod \"network-metrics-daemon-rs5rq\" (UID: \"51b171e4-cc29-4118-908c-d13825362e1c\") " pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.676774 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.694131 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.706879 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.719487 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.732982 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.740353 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.740379 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.740387 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.740401 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.740409 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:16Z","lastTransitionTime":"2025-12-03T10:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.750197 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.762210 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.782529 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:13.933771 5966 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.933938 5966 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934286 5966 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934436 5966 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934859 5966 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.935138 5966 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.968608 5966 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1203 10:54:13.968680 5966 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1203 10:54:13.968829 5966 ovnkube.go:599] Stopped ovnkube\\\\nI1203 10:54:13.968859 5966 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 10:54:13.968971 5966 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.800409 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:16Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.843013 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.843048 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.843060 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.843075 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.843086 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:16Z","lastTransitionTime":"2025-12-03T10:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.847194 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:16 crc kubenswrapper[4646]: E1203 10:54:16.847426 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.945567 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.945613 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.945625 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.945643 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:16 crc kubenswrapper[4646]: I1203 10:54:16.945656 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:16Z","lastTransitionTime":"2025-12-03T10:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.048290 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.048384 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.048409 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.048437 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.048461 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:17Z","lastTransitionTime":"2025-12-03T10:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.151369 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.151436 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.151447 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.151465 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.151480 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:17Z","lastTransitionTime":"2025-12-03T10:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.164752 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs\") pod \"network-metrics-daemon-rs5rq\" (UID: \"51b171e4-cc29-4118-908c-d13825362e1c\") " pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.164929 4646 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.164979 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs podName:51b171e4-cc29-4118-908c-d13825362e1c nodeName:}" failed. No retries permitted until 2025-12-03 10:54:18.164964035 +0000 UTC m=+34.628020180 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs") pod "network-metrics-daemon-rs5rq" (UID: "51b171e4-cc29-4118-908c-d13825362e1c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.254133 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.254189 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.254213 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.254241 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.254259 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:17Z","lastTransitionTime":"2025-12-03T10:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.357573 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.357630 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.357650 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.357679 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.357699 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:17Z","lastTransitionTime":"2025-12-03T10:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.461468 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.461791 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.461808 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.461830 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.461848 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:17Z","lastTransitionTime":"2025-12-03T10:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.565781 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.566155 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.566283 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.566440 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.566514 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:17Z","lastTransitionTime":"2025-12-03T10:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.669019 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.669204 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.669222 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:54:33.669180087 +0000 UTC m=+50.132236282 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.669282 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.669384 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.669474 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.669485 4646 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.669622 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.669648 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:33.66961638 +0000 UTC m=+50.132672555 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.669650 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.669695 4646 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.669733 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.669756 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:33.669736783 +0000 UTC m=+50.132792968 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.669759 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.669795 4646 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.669869 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:33.669846366 +0000 UTC m=+50.132902541 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.669866 4646 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.670037 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:54:33.66996583 +0000 UTC m=+50.133022015 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.670453 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.670496 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.670505 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.670521 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.670531 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:17Z","lastTransitionTime":"2025-12-03T10:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.773267 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.773324 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.773353 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.773379 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.773393 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:17Z","lastTransitionTime":"2025-12-03T10:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.848370 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.848377 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.848565 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.848652 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.849089 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:17 crc kubenswrapper[4646]: E1203 10:54:17.849231 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.877391 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.877449 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.877465 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.877487 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.877510 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:17Z","lastTransitionTime":"2025-12-03T10:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.980803 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.980846 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.980856 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.980871 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:17 crc kubenswrapper[4646]: I1203 10:54:17.980880 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:17Z","lastTransitionTime":"2025-12-03T10:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.084066 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.084110 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.084120 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.084137 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.084148 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:18Z","lastTransitionTime":"2025-12-03T10:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.174064 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs\") pod \"network-metrics-daemon-rs5rq\" (UID: \"51b171e4-cc29-4118-908c-d13825362e1c\") " pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:18 crc kubenswrapper[4646]: E1203 10:54:18.174208 4646 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:54:18 crc kubenswrapper[4646]: E1203 10:54:18.174318 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs podName:51b171e4-cc29-4118-908c-d13825362e1c nodeName:}" failed. No retries permitted until 2025-12-03 10:54:20.174300675 +0000 UTC m=+36.637356810 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs") pod "network-metrics-daemon-rs5rq" (UID: "51b171e4-cc29-4118-908c-d13825362e1c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.186468 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.186527 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.186543 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.186565 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.186577 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:18Z","lastTransitionTime":"2025-12-03T10:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.289559 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.289612 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.289633 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.289659 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.289679 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:18Z","lastTransitionTime":"2025-12-03T10:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.392708 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.392772 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.392786 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.392811 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.392826 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:18Z","lastTransitionTime":"2025-12-03T10:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.494630 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.494659 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.494667 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.494681 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.494689 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:18Z","lastTransitionTime":"2025-12-03T10:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.596845 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.596891 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.596910 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.596934 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.596952 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:18Z","lastTransitionTime":"2025-12-03T10:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.700078 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.700120 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.700131 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.700149 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.700161 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:18Z","lastTransitionTime":"2025-12-03T10:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.802890 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.802940 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.802953 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.802973 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.802985 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:18Z","lastTransitionTime":"2025-12-03T10:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.847530 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:18 crc kubenswrapper[4646]: E1203 10:54:18.847692 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.905968 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.906028 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.906045 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.906084 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:18 crc kubenswrapper[4646]: I1203 10:54:18.906103 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:18Z","lastTransitionTime":"2025-12-03T10:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.009138 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.009194 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.009212 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.009235 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.009253 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:19Z","lastTransitionTime":"2025-12-03T10:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.111617 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.111680 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.111697 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.111725 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.111742 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:19Z","lastTransitionTime":"2025-12-03T10:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.215020 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.215095 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.215125 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.215155 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.215182 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:19Z","lastTransitionTime":"2025-12-03T10:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.317689 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.317758 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.317796 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.317833 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.317852 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:19Z","lastTransitionTime":"2025-12-03T10:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.420189 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.420218 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.420227 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.420239 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.420248 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:19Z","lastTransitionTime":"2025-12-03T10:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.522874 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.522945 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.522970 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.523007 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.523030 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:19Z","lastTransitionTime":"2025-12-03T10:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.626762 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.626830 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.626846 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.626871 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.626889 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:19Z","lastTransitionTime":"2025-12-03T10:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.730184 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.730241 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.730261 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.730287 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.730304 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:19Z","lastTransitionTime":"2025-12-03T10:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.832734 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.832803 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.832827 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.832857 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.832877 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:19Z","lastTransitionTime":"2025-12-03T10:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.848224 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.848226 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.848260 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:19 crc kubenswrapper[4646]: E1203 10:54:19.848499 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:19 crc kubenswrapper[4646]: E1203 10:54:19.848580 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:19 crc kubenswrapper[4646]: E1203 10:54:19.848648 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.941675 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.941759 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.941783 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.941813 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:19 crc kubenswrapper[4646]: I1203 10:54:19.941837 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:19Z","lastTransitionTime":"2025-12-03T10:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.044541 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.044587 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.044603 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.044625 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.044641 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:20Z","lastTransitionTime":"2025-12-03T10:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.147407 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.147472 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.147490 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.147515 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.147532 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:20Z","lastTransitionTime":"2025-12-03T10:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.197363 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs\") pod \"network-metrics-daemon-rs5rq\" (UID: \"51b171e4-cc29-4118-908c-d13825362e1c\") " pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:20 crc kubenswrapper[4646]: E1203 10:54:20.197498 4646 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:54:20 crc kubenswrapper[4646]: E1203 10:54:20.197562 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs podName:51b171e4-cc29-4118-908c-d13825362e1c nodeName:}" failed. No retries permitted until 2025-12-03 10:54:24.197545007 +0000 UTC m=+40.660601142 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs") pod "network-metrics-daemon-rs5rq" (UID: "51b171e4-cc29-4118-908c-d13825362e1c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.250767 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.250814 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.250826 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.250842 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.250871 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:20Z","lastTransitionTime":"2025-12-03T10:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.354030 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.354086 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.354124 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.354148 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.354164 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:20Z","lastTransitionTime":"2025-12-03T10:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.457018 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.457050 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.457061 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.457076 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.457109 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:20Z","lastTransitionTime":"2025-12-03T10:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.560171 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.560228 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.560252 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.560279 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.560306 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:20Z","lastTransitionTime":"2025-12-03T10:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.663524 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.663581 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.663603 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.663630 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.663651 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:20Z","lastTransitionTime":"2025-12-03T10:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.766190 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.766238 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.766250 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.766267 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.766279 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:20Z","lastTransitionTime":"2025-12-03T10:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.848515 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:20 crc kubenswrapper[4646]: E1203 10:54:20.848645 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.868575 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.868638 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.868649 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.868668 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.868680 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:20Z","lastTransitionTime":"2025-12-03T10:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.971500 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.971593 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.971607 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.971624 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:20 crc kubenswrapper[4646]: I1203 10:54:20.971648 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:20Z","lastTransitionTime":"2025-12-03T10:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.074327 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.074406 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.074423 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.074478 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.074500 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:21Z","lastTransitionTime":"2025-12-03T10:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.177609 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.177657 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.177672 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.177694 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.177709 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:21Z","lastTransitionTime":"2025-12-03T10:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.280470 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.280498 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.280506 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.280547 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.280564 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:21Z","lastTransitionTime":"2025-12-03T10:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.383753 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.383839 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.383871 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.383896 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.383912 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:21Z","lastTransitionTime":"2025-12-03T10:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.487389 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.487471 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.487493 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.488028 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.488126 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:21Z","lastTransitionTime":"2025-12-03T10:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.591390 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.591434 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.591453 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.591477 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.591495 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:21Z","lastTransitionTime":"2025-12-03T10:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.695242 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.695294 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.695311 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.695376 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.695401 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:21Z","lastTransitionTime":"2025-12-03T10:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.797603 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.797668 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.797681 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.797698 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.797707 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:21Z","lastTransitionTime":"2025-12-03T10:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.848029 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.848105 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.848029 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:21 crc kubenswrapper[4646]: E1203 10:54:21.848176 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:21 crc kubenswrapper[4646]: E1203 10:54:21.848463 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:21 crc kubenswrapper[4646]: E1203 10:54:21.848326 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.900042 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.900091 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.900103 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.900121 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:21 crc kubenswrapper[4646]: I1203 10:54:21.900133 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:21Z","lastTransitionTime":"2025-12-03T10:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.002620 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.002662 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.002674 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.002688 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.002699 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:22Z","lastTransitionTime":"2025-12-03T10:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.105030 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.105070 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.105086 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.105139 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.105162 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:22Z","lastTransitionTime":"2025-12-03T10:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.208552 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.208604 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.208622 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.208656 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.208673 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:22Z","lastTransitionTime":"2025-12-03T10:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.312009 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.312062 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.312089 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.312114 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.312132 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:22Z","lastTransitionTime":"2025-12-03T10:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.414893 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.414922 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.414930 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.414944 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.414952 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:22Z","lastTransitionTime":"2025-12-03T10:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.518307 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.518432 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.518458 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.518491 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.518515 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:22Z","lastTransitionTime":"2025-12-03T10:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.621684 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.621719 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.621727 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.621741 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.621750 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:22Z","lastTransitionTime":"2025-12-03T10:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.723258 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.723295 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.723306 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.723321 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.723348 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:22Z","lastTransitionTime":"2025-12-03T10:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.825555 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.825590 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.825600 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.825615 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.825624 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:22Z","lastTransitionTime":"2025-12-03T10:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.848003 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:22 crc kubenswrapper[4646]: E1203 10:54:22.848121 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.928710 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.928759 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.928773 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.928794 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:22 crc kubenswrapper[4646]: I1203 10:54:22.928811 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:22Z","lastTransitionTime":"2025-12-03T10:54:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.030389 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.030475 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.030506 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.030521 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.030530 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:23Z","lastTransitionTime":"2025-12-03T10:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.132545 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.132590 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.132602 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.132621 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.132633 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:23Z","lastTransitionTime":"2025-12-03T10:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.234992 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.235019 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.235028 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.235041 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.235050 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:23Z","lastTransitionTime":"2025-12-03T10:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.337654 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.337728 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.337750 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.337780 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.337803 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:23Z","lastTransitionTime":"2025-12-03T10:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.440325 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.440401 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.440415 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.440432 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.440447 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:23Z","lastTransitionTime":"2025-12-03T10:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.543555 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.543621 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.543639 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.543669 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.543687 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:23Z","lastTransitionTime":"2025-12-03T10:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.646737 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.646781 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.646797 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.646837 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.646853 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:23Z","lastTransitionTime":"2025-12-03T10:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.749239 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.749305 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.749326 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.749389 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.749412 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:23Z","lastTransitionTime":"2025-12-03T10:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.848500 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.848664 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:23 crc kubenswrapper[4646]: E1203 10:54:23.848797 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.848848 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:23 crc kubenswrapper[4646]: E1203 10:54:23.849044 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:23 crc kubenswrapper[4646]: E1203 10:54:23.849122 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.854105 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.854179 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.854215 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.854265 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.854287 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:23Z","lastTransitionTime":"2025-12-03T10:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.877381 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:23Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.893260 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:23Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.906048 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:23Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.921054 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:23Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.937279 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:23Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.950912 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:23Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.956838 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.956875 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.956886 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.956903 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.956914 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:23Z","lastTransitionTime":"2025-12-03T10:54:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.977525 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:13.933771 5966 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.933938 5966 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934286 5966 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934436 5966 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934859 5966 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.935138 5966 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.968608 5966 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1203 10:54:13.968680 5966 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1203 10:54:13.968829 5966 ovnkube.go:599] Stopped ovnkube\\\\nI1203 10:54:13.968859 5966 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 10:54:13.968971 5966 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:23Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:23 crc kubenswrapper[4646]: I1203 10:54:23.989602 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:23Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.000931 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:23Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.013890 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b171e4-cc29-4118-908c-d13825362e1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:24Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.024890 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:24Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.035933 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:24Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.049064 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:24Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.058351 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:24Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.059062 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.059143 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.059197 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.059256 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.059325 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:24Z","lastTransitionTime":"2025-12-03T10:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.068477 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:24Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.076643 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:24Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.161582 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.161651 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.161668 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.161690 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.161705 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:24Z","lastTransitionTime":"2025-12-03T10:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.236003 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs\") pod \"network-metrics-daemon-rs5rq\" (UID: \"51b171e4-cc29-4118-908c-d13825362e1c\") " pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:24 crc kubenswrapper[4646]: E1203 10:54:24.236247 4646 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:54:24 crc kubenswrapper[4646]: E1203 10:54:24.236328 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs podName:51b171e4-cc29-4118-908c-d13825362e1c nodeName:}" failed. No retries permitted until 2025-12-03 10:54:32.236303067 +0000 UTC m=+48.699359242 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs") pod "network-metrics-daemon-rs5rq" (UID: "51b171e4-cc29-4118-908c-d13825362e1c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.265875 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.265939 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.265961 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.265989 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.266011 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:24Z","lastTransitionTime":"2025-12-03T10:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.368619 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.368685 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.368707 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.368734 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.368758 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:24Z","lastTransitionTime":"2025-12-03T10:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.472106 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.472170 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.472193 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.472219 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.472236 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:24Z","lastTransitionTime":"2025-12-03T10:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.574721 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.574771 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.574793 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.574817 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.574835 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:24Z","lastTransitionTime":"2025-12-03T10:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.678470 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.678819 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.678940 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.679095 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.679208 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:24Z","lastTransitionTime":"2025-12-03T10:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.790307 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.790764 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.790982 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.791216 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.791486 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:24Z","lastTransitionTime":"2025-12-03T10:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.848149 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:24 crc kubenswrapper[4646]: E1203 10:54:24.848674 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.894857 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.895187 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.895506 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.895718 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.895869 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:24Z","lastTransitionTime":"2025-12-03T10:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.998292 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.998700 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.998854 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.999112 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:24 crc kubenswrapper[4646]: I1203 10:54:24.999273 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:24Z","lastTransitionTime":"2025-12-03T10:54:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.037313 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.037683 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.037813 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.037944 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.038130 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:25Z","lastTransitionTime":"2025-12-03T10:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:25 crc kubenswrapper[4646]: E1203 10:54:25.061884 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:25Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.066320 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.066562 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.066676 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.066770 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.066855 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:25Z","lastTransitionTime":"2025-12-03T10:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:25 crc kubenswrapper[4646]: E1203 10:54:25.080866 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:25Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.084977 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.085174 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.085432 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.085536 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.085794 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:25Z","lastTransitionTime":"2025-12-03T10:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:25 crc kubenswrapper[4646]: E1203 10:54:25.104254 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:25Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.107949 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.107996 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.108014 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.108035 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.108048 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:25Z","lastTransitionTime":"2025-12-03T10:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:25 crc kubenswrapper[4646]: E1203 10:54:25.122102 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:25Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.125776 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.125822 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.125836 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.125855 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.125867 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:25Z","lastTransitionTime":"2025-12-03T10:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:25 crc kubenswrapper[4646]: E1203 10:54:25.138059 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:25Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:25 crc kubenswrapper[4646]: E1203 10:54:25.138611 4646 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.140152 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.140179 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.140188 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.140201 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.140210 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:25Z","lastTransitionTime":"2025-12-03T10:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.242946 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.242997 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.243009 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.243026 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.243037 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:25Z","lastTransitionTime":"2025-12-03T10:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.346399 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.346893 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.347195 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.347305 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.347546 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:25Z","lastTransitionTime":"2025-12-03T10:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.450897 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.450920 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.450928 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.450942 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.450951 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:25Z","lastTransitionTime":"2025-12-03T10:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.553220 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.553632 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.553796 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.554412 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.554438 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:25Z","lastTransitionTime":"2025-12-03T10:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.656649 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.657209 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.657515 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.658139 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.658613 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:25Z","lastTransitionTime":"2025-12-03T10:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.761914 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.762303 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.762740 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.762912 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.763285 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:25Z","lastTransitionTime":"2025-12-03T10:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.848067 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:25 crc kubenswrapper[4646]: E1203 10:54:25.848632 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.848452 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:25 crc kubenswrapper[4646]: E1203 10:54:25.849644 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.848432 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:25 crc kubenswrapper[4646]: E1203 10:54:25.850116 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.866297 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.866397 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.866421 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.866466 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.866496 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:25Z","lastTransitionTime":"2025-12-03T10:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.970198 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.970278 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.970299 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.970324 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:25 crc kubenswrapper[4646]: I1203 10:54:25.970372 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:25Z","lastTransitionTime":"2025-12-03T10:54:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.072724 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.073083 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.073291 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.073780 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.074168 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:26Z","lastTransitionTime":"2025-12-03T10:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.177805 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.177852 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.177865 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.177882 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.177894 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:26Z","lastTransitionTime":"2025-12-03T10:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.281567 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.281628 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.281644 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.281669 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.281686 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:26Z","lastTransitionTime":"2025-12-03T10:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.384410 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.384454 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.384471 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.384499 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.384518 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:26Z","lastTransitionTime":"2025-12-03T10:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.487381 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.487457 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.487475 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.487501 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.487518 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:26Z","lastTransitionTime":"2025-12-03T10:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.589594 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.589945 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.590186 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.590314 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.590607 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:26Z","lastTransitionTime":"2025-12-03T10:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.693572 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.693625 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.693643 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.693667 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.693684 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:26Z","lastTransitionTime":"2025-12-03T10:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.796233 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.796266 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.796277 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.796291 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.796301 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:26Z","lastTransitionTime":"2025-12-03T10:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.847531 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:26 crc kubenswrapper[4646]: E1203 10:54:26.847680 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.899823 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.900222 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.900481 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.900684 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:26 crc kubenswrapper[4646]: I1203 10:54:26.900873 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:26Z","lastTransitionTime":"2025-12-03T10:54:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.005238 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.005839 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.006143 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.006492 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.006734 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:27Z","lastTransitionTime":"2025-12-03T10:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.111714 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.111762 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.111779 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.111800 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.111813 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:27Z","lastTransitionTime":"2025-12-03T10:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.215087 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.215131 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.215142 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.215162 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.215176 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:27Z","lastTransitionTime":"2025-12-03T10:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.318146 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.318188 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.318199 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.318221 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.318234 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:27Z","lastTransitionTime":"2025-12-03T10:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.421661 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.421721 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.421740 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.421764 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.421783 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:27Z","lastTransitionTime":"2025-12-03T10:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.524450 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.524505 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.524516 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.524542 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.524557 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:27Z","lastTransitionTime":"2025-12-03T10:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.627888 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.628217 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.628308 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.628484 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.628602 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:27Z","lastTransitionTime":"2025-12-03T10:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.732071 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.732152 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.732180 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.732215 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.732239 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:27Z","lastTransitionTime":"2025-12-03T10:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.835593 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.835675 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.835700 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.835732 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.835756 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:27Z","lastTransitionTime":"2025-12-03T10:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.848114 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.848274 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.848132 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:27 crc kubenswrapper[4646]: E1203 10:54:27.848445 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:27 crc kubenswrapper[4646]: E1203 10:54:27.848597 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:27 crc kubenswrapper[4646]: E1203 10:54:27.848756 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.940268 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.940365 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.940407 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.940451 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:27 crc kubenswrapper[4646]: I1203 10:54:27.940475 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:27Z","lastTransitionTime":"2025-12-03T10:54:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.044233 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.044302 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.044321 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.044401 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.044436 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:28Z","lastTransitionTime":"2025-12-03T10:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.147250 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.147620 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.147644 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.148024 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.148324 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:28Z","lastTransitionTime":"2025-12-03T10:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.252267 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.252369 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.252391 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.252421 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.252447 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:28Z","lastTransitionTime":"2025-12-03T10:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.356158 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.356259 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.356314 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.356357 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.356376 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:28Z","lastTransitionTime":"2025-12-03T10:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.460308 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.460409 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.460430 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.461019 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.461095 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:28Z","lastTransitionTime":"2025-12-03T10:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.565282 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.565394 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.565414 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.565438 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.565460 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:28Z","lastTransitionTime":"2025-12-03T10:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.669430 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.669487 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.669672 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.669704 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.669722 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:28Z","lastTransitionTime":"2025-12-03T10:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.773056 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.773477 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.773726 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.773931 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.774101 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:28Z","lastTransitionTime":"2025-12-03T10:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.847277 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:28 crc kubenswrapper[4646]: E1203 10:54:28.847583 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.877783 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.877862 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.877888 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.877924 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.877948 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:28Z","lastTransitionTime":"2025-12-03T10:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.981254 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.981313 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.981329 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.981405 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:28 crc kubenswrapper[4646]: I1203 10:54:28.981446 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:28Z","lastTransitionTime":"2025-12-03T10:54:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.084636 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.084683 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.084695 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.084718 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.084734 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:29Z","lastTransitionTime":"2025-12-03T10:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.188260 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.188317 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.188376 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.188409 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.188432 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:29Z","lastTransitionTime":"2025-12-03T10:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.297814 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.297886 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.297908 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.297937 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.297959 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:29Z","lastTransitionTime":"2025-12-03T10:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.400626 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.400692 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.400716 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.400746 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.400767 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:29Z","lastTransitionTime":"2025-12-03T10:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.503559 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.503621 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.503639 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.503662 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.503682 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:29Z","lastTransitionTime":"2025-12-03T10:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.605888 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.605984 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.605996 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.606010 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.606018 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:29Z","lastTransitionTime":"2025-12-03T10:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.708804 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.708976 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.708997 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.709022 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.709041 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:29Z","lastTransitionTime":"2025-12-03T10:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.811909 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.812257 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.812512 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.812740 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.812914 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:29Z","lastTransitionTime":"2025-12-03T10:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.847781 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:29 crc kubenswrapper[4646]: E1203 10:54:29.847942 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.848484 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:29 crc kubenswrapper[4646]: E1203 10:54:29.848591 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.848765 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:29 crc kubenswrapper[4646]: E1203 10:54:29.848858 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.916328 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.916718 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.916845 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.917010 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:29 crc kubenswrapper[4646]: I1203 10:54:29.917134 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:29Z","lastTransitionTime":"2025-12-03T10:54:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.020774 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.020853 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.020880 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.020910 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.020933 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:30Z","lastTransitionTime":"2025-12-03T10:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.124207 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.124478 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.124509 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.124536 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.124558 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:30Z","lastTransitionTime":"2025-12-03T10:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.228060 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.228114 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.228131 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.228155 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.228172 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:30Z","lastTransitionTime":"2025-12-03T10:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.331488 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.331543 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.331559 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.331583 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.331599 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:30Z","lastTransitionTime":"2025-12-03T10:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.434828 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.434889 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.434911 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.434940 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.434962 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:30Z","lastTransitionTime":"2025-12-03T10:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.537893 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.537959 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.537978 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.538003 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.538018 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:30Z","lastTransitionTime":"2025-12-03T10:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.640479 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.640542 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.640560 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.640585 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.640603 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:30Z","lastTransitionTime":"2025-12-03T10:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.743966 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.744029 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.744056 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.744086 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.744108 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:30Z","lastTransitionTime":"2025-12-03T10:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.846749 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.847041 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.847201 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.847243 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.847380 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.847482 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:30Z","lastTransitionTime":"2025-12-03T10:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:30 crc kubenswrapper[4646]: E1203 10:54:30.848050 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.848555 4646 scope.go:117] "RemoveContainer" containerID="4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.951330 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.951692 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.951713 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.951740 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:30 crc kubenswrapper[4646]: I1203 10:54:30.951757 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:30Z","lastTransitionTime":"2025-12-03T10:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.055418 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.055467 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.055478 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.055495 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.055506 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:31Z","lastTransitionTime":"2025-12-03T10:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.159050 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.159081 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.159092 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.159109 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.159121 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:31Z","lastTransitionTime":"2025-12-03T10:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.159875 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovnkube-controller/1.log" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.164979 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerStarted","Data":"bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3"} Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.167415 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.180255 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.199595 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.215400 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.238664 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.251713 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.265377 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.288770 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.288811 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.288824 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.288840 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.288857 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:31Z","lastTransitionTime":"2025-12-03T10:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.290536 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.311861 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.336620 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.349482 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.370979 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:13.933771 5966 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.933938 5966 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934286 5966 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934436 5966 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934859 5966 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.935138 5966 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.968608 5966 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1203 10:54:13.968680 5966 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1203 10:54:13.968829 5966 ovnkube.go:599] Stopped ovnkube\\\\nI1203 10:54:13.968859 5966 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 10:54:13.968971 5966 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.386759 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.391001 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.391036 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.391045 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.391061 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.391070 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:31Z","lastTransitionTime":"2025-12-03T10:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.399871 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.411219 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.422553 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.433076 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b171e4-cc29-4118-908c-d13825362e1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:31Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.492590 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.492635 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.492651 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.492669 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.492680 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:31Z","lastTransitionTime":"2025-12-03T10:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.595089 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.595129 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.595143 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.595164 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.595179 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:31Z","lastTransitionTime":"2025-12-03T10:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.698244 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.698290 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.698302 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.698320 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.698348 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:31Z","lastTransitionTime":"2025-12-03T10:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.801562 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.801606 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.801622 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.801640 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.801652 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:31Z","lastTransitionTime":"2025-12-03T10:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.848120 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:31 crc kubenswrapper[4646]: E1203 10:54:31.848264 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.848444 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.848645 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:31 crc kubenswrapper[4646]: E1203 10:54:31.848720 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:31 crc kubenswrapper[4646]: E1203 10:54:31.848638 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.904695 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.904760 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.904784 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.904816 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:31 crc kubenswrapper[4646]: I1203 10:54:31.904839 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:31Z","lastTransitionTime":"2025-12-03T10:54:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.008282 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.008396 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.008417 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.008443 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.008461 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:32Z","lastTransitionTime":"2025-12-03T10:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.111454 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.111502 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.111514 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.111535 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.111548 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:32Z","lastTransitionTime":"2025-12-03T10:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.172641 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovnkube-controller/2.log" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.173876 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovnkube-controller/1.log" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.178821 4646 generic.go:334] "Generic (PLEG): container finished" podID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerID="bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3" exitCode=1 Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.178880 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerDied","Data":"bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3"} Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.178955 4646 scope.go:117] "RemoveContainer" containerID="4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.180464 4646 scope.go:117] "RemoveContainer" containerID="bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3" Dec 03 10:54:32 crc kubenswrapper[4646]: E1203 10:54:32.180775 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.200519 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.214613 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.214652 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.214664 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.214682 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.214695 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:32Z","lastTransitionTime":"2025-12-03T10:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.217585 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.232586 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.247991 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.266312 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.291470 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4da52703175381366d498d47ce1ce52ff4a40073ebe74812dab0eb4b037effff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"message\\\":\\\"0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:13.933771 5966 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.933938 5966 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934286 5966 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934436 5966 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.934859 5966 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.935138 5966 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:13.968608 5966 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1203 10:54:13.968680 5966 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1203 10:54:13.968829 5966 ovnkube.go:599] Stopped ovnkube\\\\nI1203 10:54:13.968859 5966 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 10:54:13.968971 5966 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:31Z\\\",\\\"message\\\":\\\"o-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:31.771666 6174 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:31.772446 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 10:54:31.772522 6174 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:31.772536 6174 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:31.772544 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:31.772569 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:31.772589 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:31.772615 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:31.772621 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:31.772647 6174 factory.go:656] Stopping watch factory\\\\nI1203 10:54:31.772671 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 10:54:31.772680 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:31.772688 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:31.772703 6174 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.318721 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.322665 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.322741 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.322754 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.322773 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.322786 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:32Z","lastTransitionTime":"2025-12-03T10:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.333404 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs\") pod \"network-metrics-daemon-rs5rq\" (UID: \"51b171e4-cc29-4118-908c-d13825362e1c\") " pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:32 crc kubenswrapper[4646]: E1203 10:54:32.333773 4646 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:54:32 crc kubenswrapper[4646]: E1203 10:54:32.333880 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs podName:51b171e4-cc29-4118-908c-d13825362e1c nodeName:}" failed. No retries permitted until 2025-12-03 10:54:48.333855377 +0000 UTC m=+64.796911522 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs") pod "network-metrics-daemon-rs5rq" (UID: "51b171e4-cc29-4118-908c-d13825362e1c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.337249 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.353997 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.370432 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b171e4-cc29-4118-908c-d13825362e1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.386761 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.401299 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.414775 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.425114 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.426534 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.426602 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.426616 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.426637 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.426649 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:32Z","lastTransitionTime":"2025-12-03T10:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.439528 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.451468 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:32Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.528651 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.528704 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.528722 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.528748 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.528766 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:32Z","lastTransitionTime":"2025-12-03T10:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.632280 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.632373 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.632395 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.632421 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.632442 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:32Z","lastTransitionTime":"2025-12-03T10:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.735867 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.735916 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.735926 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.735945 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.735960 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:32Z","lastTransitionTime":"2025-12-03T10:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.839491 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.839599 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.839622 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.839707 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.839733 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:32Z","lastTransitionTime":"2025-12-03T10:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.848261 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:32 crc kubenswrapper[4646]: E1203 10:54:32.848450 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.942835 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.942913 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.942928 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.942960 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:32 crc kubenswrapper[4646]: I1203 10:54:32.942980 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:32Z","lastTransitionTime":"2025-12-03T10:54:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.046595 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.046959 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.047200 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.047416 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.047559 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:33Z","lastTransitionTime":"2025-12-03T10:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.150531 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.150930 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.151108 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.151267 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.151451 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:33Z","lastTransitionTime":"2025-12-03T10:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.187085 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovnkube-controller/2.log" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.192685 4646 scope.go:117] "RemoveContainer" containerID="bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3" Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.192940 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.211019 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.228073 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b171e4-cc29-4118-908c-d13825362e1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.246481 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.256796 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.256860 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.256873 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.256896 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.256913 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:33Z","lastTransitionTime":"2025-12-03T10:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.271504 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.288315 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.311040 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.327036 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.345793 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.352327 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.360495 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.360547 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.360559 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.360583 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.360597 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:33Z","lastTransitionTime":"2025-12-03T10:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.365832 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.376528 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.398448 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.418388 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.440559 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.457775 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.465277 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.465400 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.465519 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.465582 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.465610 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:33Z","lastTransitionTime":"2025-12-03T10:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.490936 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:31Z\\\",\\\"message\\\":\\\"o-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:31.771666 6174 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:31.772446 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 10:54:31.772522 6174 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:31.772536 6174 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:31.772544 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:31.772569 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:31.772589 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:31.772615 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:31.772621 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:31.772647 6174 factory.go:656] Stopping watch factory\\\\nI1203 10:54:31.772671 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 10:54:31.772680 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:31.772688 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:31.772703 6174 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.515041 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.538262 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.590412 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.590663 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.590741 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.590825 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.590906 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:33Z","lastTransitionTime":"2025-12-03T10:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.604050 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.618064 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.629995 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b171e4-cc29-4118-908c-d13825362e1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.651057 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.670476 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.685804 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.688882 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.689007 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.689047 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.689134 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:55:05.689097008 +0000 UTC m=+82.152153183 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.689190 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.689219 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.689228 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.689324 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.689235 4646 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.689446 4646 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.689493 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:55:05.689469459 +0000 UTC m=+82.152525634 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.689240 4646 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.689547 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:55:05.68952495 +0000 UTC m=+82.152581125 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.689396 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.689585 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.689603 4646 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.689585 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 10:55:05.689563261 +0000 UTC m=+82.152619436 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.689667 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 10:55:05.689653924 +0000 UTC m=+82.152710099 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.693315 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.693519 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.693645 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.693740 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.693834 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:33Z","lastTransitionTime":"2025-12-03T10:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.702559 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.721180 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.735022 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.752434 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.772230 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.787855 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e90223ba-3577-408c-a674-92a9aaebc640\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8cbb08ab6f4da24ce8dcbafd255b9dd3461bc80abbd6d750e34abcbac41bb71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://821dd6864300c8286093c2ac960632ba03ed444f90994aa9cc24080e93472a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30f13d59d6c9c0046b62a4e9b202ba22e20014327ebd11dc59dc871a318826c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.795785 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.795821 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.795833 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.795850 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.795864 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:33Z","lastTransitionTime":"2025-12-03T10:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.807529 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.826510 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.842965 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.847943 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.848255 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.847992 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.848536 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.847979 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:33 crc kubenswrapper[4646]: E1203 10:54:33.848861 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.867664 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:31Z\\\",\\\"message\\\":\\\"o-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:31.771666 6174 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:31.772446 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 10:54:31.772522 6174 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:31.772536 6174 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:31.772544 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:31.772569 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:31.772589 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:31.772615 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:31.772621 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:31.772647 6174 factory.go:656] Stopping watch factory\\\\nI1203 10:54:31.772671 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 10:54:31.772680 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:31.772688 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:31.772703 6174 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.889172 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.899583 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.899632 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.899650 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.899676 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.899692 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:33Z","lastTransitionTime":"2025-12-03T10:54:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.906302 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.924288 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.940530 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b171e4-cc29-4118-908c-d13825362e1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.960558 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.980300 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:33 crc kubenswrapper[4646]: I1203 10:54:33.995606 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:33Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.004865 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.005005 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.005085 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.005170 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.005258 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:34Z","lastTransitionTime":"2025-12-03T10:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.009484 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:34Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.027504 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:34Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.039696 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:34Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.059883 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:34Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.080126 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:34Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.102016 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e90223ba-3577-408c-a674-92a9aaebc640\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8cbb08ab6f4da24ce8dcbafd255b9dd3461bc80abbd6d750e34abcbac41bb71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://821dd6864300c8286093c2ac960632ba03ed444f90994aa9cc24080e93472a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30f13d59d6c9c0046b62a4e9b202ba22e20014327ebd11dc59dc871a318826c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:34Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.107399 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.107578 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.107690 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.107821 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.107926 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:34Z","lastTransitionTime":"2025-12-03T10:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.118831 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:34Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.139955 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:34Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.156236 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:34Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.181967 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:31Z\\\",\\\"message\\\":\\\"o-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:31.771666 6174 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:31.772446 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 10:54:31.772522 6174 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:31.772536 6174 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:31.772544 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:31.772569 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:31.772589 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:31.772615 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:31.772621 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:31.772647 6174 factory.go:656] Stopping watch factory\\\\nI1203 10:54:31.772671 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 10:54:31.772680 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:31.772688 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:31.772703 6174 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:34Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.209413 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:34Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.210694 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.210729 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.210740 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.210758 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.210770 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:34Z","lastTransitionTime":"2025-12-03T10:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.313394 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.313433 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.313445 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.313485 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.313497 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:34Z","lastTransitionTime":"2025-12-03T10:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.415888 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.415919 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.415928 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.415941 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.415950 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:34Z","lastTransitionTime":"2025-12-03T10:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.518274 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.518355 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.518371 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.518396 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.518411 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:34Z","lastTransitionTime":"2025-12-03T10:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.622010 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.622076 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.622093 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.622118 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.622137 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:34Z","lastTransitionTime":"2025-12-03T10:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.725529 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.725595 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.725611 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.725638 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.725658 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:34Z","lastTransitionTime":"2025-12-03T10:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.828649 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.828711 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.828726 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.828752 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.828771 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:34Z","lastTransitionTime":"2025-12-03T10:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.847941 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:34 crc kubenswrapper[4646]: E1203 10:54:34.848164 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.932136 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.932210 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.932236 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.932265 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:34 crc kubenswrapper[4646]: I1203 10:54:34.932286 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:34Z","lastTransitionTime":"2025-12-03T10:54:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.036094 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.036127 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.036140 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.036154 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.036162 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:35Z","lastTransitionTime":"2025-12-03T10:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.138955 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.139022 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.139038 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.139065 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.139082 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:35Z","lastTransitionTime":"2025-12-03T10:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.241914 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.242005 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.242026 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.242082 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.242099 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:35Z","lastTransitionTime":"2025-12-03T10:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.345288 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.345364 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.345384 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.345409 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.345427 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:35Z","lastTransitionTime":"2025-12-03T10:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.449080 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.449174 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.449221 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.449248 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.449301 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:35Z","lastTransitionTime":"2025-12-03T10:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.491037 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.491102 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.491121 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.491146 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.491163 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:35Z","lastTransitionTime":"2025-12-03T10:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:35 crc kubenswrapper[4646]: E1203 10:54:35.511865 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:35Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.517183 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.517239 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.517260 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.517290 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.517313 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:35Z","lastTransitionTime":"2025-12-03T10:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:35 crc kubenswrapper[4646]: E1203 10:54:35.530813 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:35Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.534879 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.534932 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.534954 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.534977 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.534992 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:35Z","lastTransitionTime":"2025-12-03T10:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:35 crc kubenswrapper[4646]: E1203 10:54:35.549867 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:35Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.554362 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.554416 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.554431 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.554452 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.554468 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:35Z","lastTransitionTime":"2025-12-03T10:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:35 crc kubenswrapper[4646]: E1203 10:54:35.568436 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:35Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.572535 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.572572 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.572584 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.572602 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.572613 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:35Z","lastTransitionTime":"2025-12-03T10:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:35 crc kubenswrapper[4646]: E1203 10:54:35.585040 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:35Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:35 crc kubenswrapper[4646]: E1203 10:54:35.585647 4646 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.588713 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.588825 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.588892 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.588927 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.589037 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:35Z","lastTransitionTime":"2025-12-03T10:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.692425 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.692490 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.692522 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.692552 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.692573 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:35Z","lastTransitionTime":"2025-12-03T10:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.795320 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.795387 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.795401 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.795419 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.795431 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:35Z","lastTransitionTime":"2025-12-03T10:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.847739 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:35 crc kubenswrapper[4646]: E1203 10:54:35.847938 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.848381 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.848490 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:35 crc kubenswrapper[4646]: E1203 10:54:35.848700 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:35 crc kubenswrapper[4646]: E1203 10:54:35.848851 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.898779 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.898833 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.898850 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.898875 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:35 crc kubenswrapper[4646]: I1203 10:54:35.898894 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:35Z","lastTransitionTime":"2025-12-03T10:54:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.001576 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.001614 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.001625 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.001641 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.001651 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:36Z","lastTransitionTime":"2025-12-03T10:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.104859 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.104907 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.104922 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.104941 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.104955 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:36Z","lastTransitionTime":"2025-12-03T10:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.206234 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.206268 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.206279 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.206294 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.206304 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:36Z","lastTransitionTime":"2025-12-03T10:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.308774 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.308812 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.308820 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.308834 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.308842 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:36Z","lastTransitionTime":"2025-12-03T10:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.411799 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.411864 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.411881 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.411905 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.411923 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:36Z","lastTransitionTime":"2025-12-03T10:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.514992 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.515061 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.515086 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.515117 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.515138 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:36Z","lastTransitionTime":"2025-12-03T10:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.618458 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.618519 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.618541 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.618570 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.618591 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:36Z","lastTransitionTime":"2025-12-03T10:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.721049 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.721118 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.721142 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.721165 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.721182 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:36Z","lastTransitionTime":"2025-12-03T10:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.823751 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.823810 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.823835 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.823865 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.823898 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:36Z","lastTransitionTime":"2025-12-03T10:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.848070 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:36 crc kubenswrapper[4646]: E1203 10:54:36.848235 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.927207 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.927282 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.927300 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.927326 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:36 crc kubenswrapper[4646]: I1203 10:54:36.927380 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:36Z","lastTransitionTime":"2025-12-03T10:54:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.029906 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.029948 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.029959 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.029976 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.029988 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:37Z","lastTransitionTime":"2025-12-03T10:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.132324 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.132390 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.132404 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.132421 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.132432 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:37Z","lastTransitionTime":"2025-12-03T10:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.234936 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.235063 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.235100 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.235130 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.235152 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:37Z","lastTransitionTime":"2025-12-03T10:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.344056 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.344149 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.344171 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.345073 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.345578 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:37Z","lastTransitionTime":"2025-12-03T10:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.448709 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.448776 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.448799 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.448828 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.448845 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:37Z","lastTransitionTime":"2025-12-03T10:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.556104 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.556164 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.556184 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.556211 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.556231 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:37Z","lastTransitionTime":"2025-12-03T10:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.659077 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.659110 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.659119 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.659135 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.659143 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:37Z","lastTransitionTime":"2025-12-03T10:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.761987 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.762034 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.762046 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.762065 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.762081 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:37Z","lastTransitionTime":"2025-12-03T10:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.848018 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.848065 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:37 crc kubenswrapper[4646]: E1203 10:54:37.848229 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.848329 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:37 crc kubenswrapper[4646]: E1203 10:54:37.848535 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:37 crc kubenswrapper[4646]: E1203 10:54:37.848660 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.864919 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.864963 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.864981 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.865005 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.865023 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:37Z","lastTransitionTime":"2025-12-03T10:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.967677 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.967727 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.967746 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.967768 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:37 crc kubenswrapper[4646]: I1203 10:54:37.967786 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:37Z","lastTransitionTime":"2025-12-03T10:54:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.233939 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.233962 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.233971 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.233985 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.233996 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:38Z","lastTransitionTime":"2025-12-03T10:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.336261 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.336599 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.336841 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.337088 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.337187 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:38Z","lastTransitionTime":"2025-12-03T10:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.440260 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.440324 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.440389 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.440417 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.440437 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:38Z","lastTransitionTime":"2025-12-03T10:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.543467 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.543525 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.543547 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.543575 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.543597 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:38Z","lastTransitionTime":"2025-12-03T10:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.646606 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.646922 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.647058 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.647196 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.647328 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:38Z","lastTransitionTime":"2025-12-03T10:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.750397 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.750677 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.750896 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.751102 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.751271 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:38Z","lastTransitionTime":"2025-12-03T10:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.848175 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:38 crc kubenswrapper[4646]: E1203 10:54:38.848445 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.854580 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.854645 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.854667 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.854697 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.854715 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:38Z","lastTransitionTime":"2025-12-03T10:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.958062 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.958116 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.958161 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.958246 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:38 crc kubenswrapper[4646]: I1203 10:54:38.958300 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:38Z","lastTransitionTime":"2025-12-03T10:54:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.060841 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.060916 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.060935 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.060960 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.060978 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:39Z","lastTransitionTime":"2025-12-03T10:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.164123 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.164187 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.164209 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.164238 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.164259 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:39Z","lastTransitionTime":"2025-12-03T10:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.267356 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.267655 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.267763 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.267856 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.267929 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:39Z","lastTransitionTime":"2025-12-03T10:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.370489 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.370852 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.371056 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.371480 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.371855 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:39Z","lastTransitionTime":"2025-12-03T10:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.475150 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.475203 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.475224 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.475248 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.475264 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:39Z","lastTransitionTime":"2025-12-03T10:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.578077 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.578126 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.578181 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.578206 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.578223 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:39Z","lastTransitionTime":"2025-12-03T10:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.682143 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.682599 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.682793 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.682996 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.683181 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:39Z","lastTransitionTime":"2025-12-03T10:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.787309 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.787381 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.787399 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.787423 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.787439 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:39Z","lastTransitionTime":"2025-12-03T10:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.848331 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.848447 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.848494 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:39 crc kubenswrapper[4646]: E1203 10:54:39.848585 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:39 crc kubenswrapper[4646]: E1203 10:54:39.848708 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:39 crc kubenswrapper[4646]: E1203 10:54:39.849241 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.891393 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.891831 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.892067 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.892287 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.892543 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:39Z","lastTransitionTime":"2025-12-03T10:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.995870 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.995921 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.995942 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.995968 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:39 crc kubenswrapper[4646]: I1203 10:54:39.995986 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:39Z","lastTransitionTime":"2025-12-03T10:54:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.098523 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.098578 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.098595 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.098619 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.098635 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:40Z","lastTransitionTime":"2025-12-03T10:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.202152 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.202208 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.202227 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.202253 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.202272 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:40Z","lastTransitionTime":"2025-12-03T10:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.304822 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.304876 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.304896 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.304923 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.304941 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:40Z","lastTransitionTime":"2025-12-03T10:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.407729 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.407774 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.407790 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.407817 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.407833 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:40Z","lastTransitionTime":"2025-12-03T10:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.510160 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.510626 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.510819 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.510983 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.511178 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:40Z","lastTransitionTime":"2025-12-03T10:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.614620 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.614989 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.615185 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.615430 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.615662 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:40Z","lastTransitionTime":"2025-12-03T10:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.719689 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.719762 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.719785 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.719814 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.719836 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:40Z","lastTransitionTime":"2025-12-03T10:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.822236 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.822301 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.822321 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.822383 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.822405 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:40Z","lastTransitionTime":"2025-12-03T10:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.847688 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:40 crc kubenswrapper[4646]: E1203 10:54:40.847906 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.925329 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.925533 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.925553 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.925578 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:40 crc kubenswrapper[4646]: I1203 10:54:40.925596 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:40Z","lastTransitionTime":"2025-12-03T10:54:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.028854 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.028909 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.028927 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.028949 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.028967 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:41Z","lastTransitionTime":"2025-12-03T10:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.131905 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.132193 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.132425 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.132517 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.132590 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:41Z","lastTransitionTime":"2025-12-03T10:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.234987 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.235445 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.235539 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.235631 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.235722 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:41Z","lastTransitionTime":"2025-12-03T10:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.338424 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.338464 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.338476 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.338495 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.338507 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:41Z","lastTransitionTime":"2025-12-03T10:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.441699 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.441758 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.441778 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.441804 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.441823 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:41Z","lastTransitionTime":"2025-12-03T10:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.545292 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.545952 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.546183 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.546517 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.546759 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:41Z","lastTransitionTime":"2025-12-03T10:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.650827 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.650890 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.650911 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.650935 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.650952 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:41Z","lastTransitionTime":"2025-12-03T10:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.753418 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.753721 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.753820 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.753924 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.754051 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:41Z","lastTransitionTime":"2025-12-03T10:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.848025 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.848080 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.848219 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:41 crc kubenswrapper[4646]: E1203 10:54:41.848930 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:41 crc kubenswrapper[4646]: E1203 10:54:41.849002 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:41 crc kubenswrapper[4646]: E1203 10:54:41.849130 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.864279 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.864377 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.864407 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.864441 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.864464 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:41Z","lastTransitionTime":"2025-12-03T10:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.967283 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.967425 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.967445 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.967471 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:41 crc kubenswrapper[4646]: I1203 10:54:41.967489 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:41Z","lastTransitionTime":"2025-12-03T10:54:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.071258 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.071739 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.071951 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.072162 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.072385 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:42Z","lastTransitionTime":"2025-12-03T10:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.176434 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.176466 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.176477 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.176495 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.176506 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:42Z","lastTransitionTime":"2025-12-03T10:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.279822 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.279903 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.279925 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.279953 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.279971 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:42Z","lastTransitionTime":"2025-12-03T10:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.384215 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.384263 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.384281 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.384306 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.384323 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:42Z","lastTransitionTime":"2025-12-03T10:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.488142 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.488306 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.488382 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.488417 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.488441 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:42Z","lastTransitionTime":"2025-12-03T10:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.591205 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.591242 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.591259 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.591282 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.591299 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:42Z","lastTransitionTime":"2025-12-03T10:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.694481 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.694538 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.694557 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.694581 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.694598 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:42Z","lastTransitionTime":"2025-12-03T10:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.798203 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.798252 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.798269 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.798293 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.798310 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:42Z","lastTransitionTime":"2025-12-03T10:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.847724 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:42 crc kubenswrapper[4646]: E1203 10:54:42.848196 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.900887 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.900965 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.900989 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.901018 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:42 crc kubenswrapper[4646]: I1203 10:54:42.901039 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:42Z","lastTransitionTime":"2025-12-03T10:54:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.003754 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.003804 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.003821 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.003846 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.003863 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:43Z","lastTransitionTime":"2025-12-03T10:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.108045 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.108460 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.108641 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.108817 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.108971 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:43Z","lastTransitionTime":"2025-12-03T10:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.211830 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.212044 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.212114 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.212229 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.212290 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:43Z","lastTransitionTime":"2025-12-03T10:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.314534 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.314782 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.315002 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.315364 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.315449 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:43Z","lastTransitionTime":"2025-12-03T10:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.418697 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.419139 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.419554 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.419610 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.419636 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:43Z","lastTransitionTime":"2025-12-03T10:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.521788 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.521836 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.521851 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.521868 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.521881 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:43Z","lastTransitionTime":"2025-12-03T10:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.624731 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.624766 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.624776 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.624799 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.624810 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:43Z","lastTransitionTime":"2025-12-03T10:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.727965 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.728022 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.728032 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.728154 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.728167 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:43Z","lastTransitionTime":"2025-12-03T10:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.831630 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.832027 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.832248 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.832643 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.832833 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:43Z","lastTransitionTime":"2025-12-03T10:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.847540 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.847619 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:43 crc kubenswrapper[4646]: E1203 10:54:43.847693 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.847619 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:43 crc kubenswrapper[4646]: E1203 10:54:43.847914 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:43 crc kubenswrapper[4646]: E1203 10:54:43.848034 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.869161 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:43Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.890097 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:43Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.914114 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:43Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.930519 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:43Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.935488 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.935536 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.935553 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.935584 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.935610 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:43Z","lastTransitionTime":"2025-12-03T10:54:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.953429 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:43Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.969292 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:43Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:43 crc kubenswrapper[4646]: I1203 10:54:43.990876 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:43Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.014732 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:44Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.030744 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:44Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.041501 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.041572 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.042185 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.042217 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.042237 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:44Z","lastTransitionTime":"2025-12-03T10:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.047322 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e90223ba-3577-408c-a674-92a9aaebc640\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8cbb08ab6f4da24ce8dcbafd255b9dd3461bc80abbd6d750e34abcbac41bb71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://821dd6864300c8286093c2ac960632ba03ed444f90994aa9cc24080e93472a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30f13d59d6c9c0046b62a4e9b202ba22e20014327ebd11dc59dc871a318826c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:44Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.063766 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:44Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.090973 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:44Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.105972 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:44Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.126680 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:31Z\\\",\\\"message\\\":\\\"o-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:31.771666 6174 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:31.772446 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 10:54:31.772522 6174 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:31.772536 6174 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:31.772544 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:31.772569 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:31.772589 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:31.772615 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:31.772621 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:31.772647 6174 factory.go:656] Stopping watch factory\\\\nI1203 10:54:31.772671 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 10:54:31.772680 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:31.772688 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:31.772703 6174 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:44Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.141541 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:44Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.143785 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.143805 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.143816 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.143831 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.143840 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:44Z","lastTransitionTime":"2025-12-03T10:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.152525 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:44Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.164203 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b171e4-cc29-4118-908c-d13825362e1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:44Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.245884 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.245915 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.245924 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.245937 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.245945 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:44Z","lastTransitionTime":"2025-12-03T10:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.349039 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.349106 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.349127 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.349153 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.349169 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:44Z","lastTransitionTime":"2025-12-03T10:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.452430 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.452462 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.452473 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.452491 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.452503 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:44Z","lastTransitionTime":"2025-12-03T10:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.556075 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.556125 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.556144 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.556170 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.556190 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:44Z","lastTransitionTime":"2025-12-03T10:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.658709 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.658769 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.658786 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.658811 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.658828 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:44Z","lastTransitionTime":"2025-12-03T10:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.761647 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.761957 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.762097 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.762245 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.762458 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:44Z","lastTransitionTime":"2025-12-03T10:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.847979 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:44 crc kubenswrapper[4646]: E1203 10:54:44.848201 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.864950 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.864985 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.864995 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.865012 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.865024 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:44Z","lastTransitionTime":"2025-12-03T10:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.968057 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.968105 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.968119 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.968139 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:44 crc kubenswrapper[4646]: I1203 10:54:44.968154 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:44Z","lastTransitionTime":"2025-12-03T10:54:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.071404 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.071459 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.071477 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.071501 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.071519 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:45Z","lastTransitionTime":"2025-12-03T10:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.174481 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.174521 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.174534 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.174551 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.174565 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:45Z","lastTransitionTime":"2025-12-03T10:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.277358 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.277409 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.277423 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.277442 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.277454 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:45Z","lastTransitionTime":"2025-12-03T10:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.380556 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.380649 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.380670 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.380697 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.380716 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:45Z","lastTransitionTime":"2025-12-03T10:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.483993 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.484045 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.484064 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.484088 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.484105 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:45Z","lastTransitionTime":"2025-12-03T10:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.586947 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.587294 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.587501 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.587687 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.587871 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:45Z","lastTransitionTime":"2025-12-03T10:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.652840 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.653157 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.653324 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.653483 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.653615 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:45Z","lastTransitionTime":"2025-12-03T10:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:45 crc kubenswrapper[4646]: E1203 10:54:45.671143 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:45Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.676369 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.676565 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.676680 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.676813 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.676932 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:45Z","lastTransitionTime":"2025-12-03T10:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:45 crc kubenswrapper[4646]: E1203 10:54:45.691837 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:45Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.697010 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.697174 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.697289 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.697425 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.697534 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:45Z","lastTransitionTime":"2025-12-03T10:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:45 crc kubenswrapper[4646]: E1203 10:54:45.713580 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:45Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.718084 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.718139 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.718275 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.718297 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.718309 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:45Z","lastTransitionTime":"2025-12-03T10:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:45 crc kubenswrapper[4646]: E1203 10:54:45.732261 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:45Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.736351 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.736392 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.736406 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.736427 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.736442 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:45Z","lastTransitionTime":"2025-12-03T10:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:45 crc kubenswrapper[4646]: E1203 10:54:45.754314 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:45Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:45 crc kubenswrapper[4646]: E1203 10:54:45.754579 4646 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.756697 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.756832 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.756936 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.757039 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.757145 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:45Z","lastTransitionTime":"2025-12-03T10:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.847932 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.848041 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:45 crc kubenswrapper[4646]: E1203 10:54:45.848177 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.847971 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:45 crc kubenswrapper[4646]: E1203 10:54:45.848739 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:45 crc kubenswrapper[4646]: E1203 10:54:45.848530 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.858904 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.859092 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.859192 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.859283 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.859395 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:45Z","lastTransitionTime":"2025-12-03T10:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.962637 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.962893 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.962985 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.963076 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:45 crc kubenswrapper[4646]: I1203 10:54:45.963153 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:45Z","lastTransitionTime":"2025-12-03T10:54:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.065926 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.066172 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.066292 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.066398 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.066489 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:46Z","lastTransitionTime":"2025-12-03T10:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.170357 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.170425 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.170447 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.170476 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.170497 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:46Z","lastTransitionTime":"2025-12-03T10:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.272529 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.272820 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.272931 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.273054 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.273167 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:46Z","lastTransitionTime":"2025-12-03T10:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.376412 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.376736 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.376844 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.376953 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.377064 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:46Z","lastTransitionTime":"2025-12-03T10:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.481682 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.482216 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.482504 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.482747 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.482979 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:46Z","lastTransitionTime":"2025-12-03T10:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.587497 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.587879 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.588065 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.588292 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.588521 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:46Z","lastTransitionTime":"2025-12-03T10:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.692280 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.692332 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.692380 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.692410 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.692431 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:46Z","lastTransitionTime":"2025-12-03T10:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.795773 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.795822 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.795842 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.795867 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.795885 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:46Z","lastTransitionTime":"2025-12-03T10:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.847581 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:46 crc kubenswrapper[4646]: E1203 10:54:46.847805 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.848898 4646 scope.go:117] "RemoveContainer" containerID="bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3" Dec 03 10:54:46 crc kubenswrapper[4646]: E1203 10:54:46.849238 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.898813 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.899018 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.899079 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.899149 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:46 crc kubenswrapper[4646]: I1203 10:54:46.899211 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:46Z","lastTransitionTime":"2025-12-03T10:54:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.001230 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.001323 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.001395 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.001423 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.001440 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:47Z","lastTransitionTime":"2025-12-03T10:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.104521 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.104583 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.104595 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.104613 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.104625 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:47Z","lastTransitionTime":"2025-12-03T10:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.207460 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.207498 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.207511 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.207527 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.207537 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:47Z","lastTransitionTime":"2025-12-03T10:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.309476 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.309526 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.309534 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.309550 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.309559 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:47Z","lastTransitionTime":"2025-12-03T10:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.411769 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.411818 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.411834 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.411856 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.411873 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:47Z","lastTransitionTime":"2025-12-03T10:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.514296 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.514864 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.515009 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.515145 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.515307 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:47Z","lastTransitionTime":"2025-12-03T10:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.617890 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.617936 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.617951 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.617972 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.617992 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:47Z","lastTransitionTime":"2025-12-03T10:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.720440 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.720498 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.720508 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.720525 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.720533 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:47Z","lastTransitionTime":"2025-12-03T10:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.822478 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.822503 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.822511 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.822525 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.822534 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:47Z","lastTransitionTime":"2025-12-03T10:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.848246 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.848246 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:47 crc kubenswrapper[4646]: E1203 10:54:47.848368 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.848396 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:47 crc kubenswrapper[4646]: E1203 10:54:47.848463 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:47 crc kubenswrapper[4646]: E1203 10:54:47.848517 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.924990 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.925283 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.925386 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.925503 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:47 crc kubenswrapper[4646]: I1203 10:54:47.925591 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:47Z","lastTransitionTime":"2025-12-03T10:54:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.029081 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.029186 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.029204 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.029229 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.029247 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:48Z","lastTransitionTime":"2025-12-03T10:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.131873 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.131909 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.131923 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.131939 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.131950 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:48Z","lastTransitionTime":"2025-12-03T10:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.234314 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.234398 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.234410 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.234428 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.234440 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:48Z","lastTransitionTime":"2025-12-03T10:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.335255 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs\") pod \"network-metrics-daemon-rs5rq\" (UID: \"51b171e4-cc29-4118-908c-d13825362e1c\") " pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:48 crc kubenswrapper[4646]: E1203 10:54:48.335467 4646 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:54:48 crc kubenswrapper[4646]: E1203 10:54:48.335573 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs podName:51b171e4-cc29-4118-908c-d13825362e1c nodeName:}" failed. No retries permitted until 2025-12-03 10:55:20.335549447 +0000 UTC m=+96.798605622 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs") pod "network-metrics-daemon-rs5rq" (UID: "51b171e4-cc29-4118-908c-d13825362e1c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.336011 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.336063 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.336075 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.336091 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.336101 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:48Z","lastTransitionTime":"2025-12-03T10:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.438083 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.438119 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.438131 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.438152 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.438165 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:48Z","lastTransitionTime":"2025-12-03T10:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.540623 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.540701 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.540723 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.540748 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.540768 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:48Z","lastTransitionTime":"2025-12-03T10:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.643100 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.643134 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.643146 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.643163 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.643174 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:48Z","lastTransitionTime":"2025-12-03T10:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.744785 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.744816 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.744825 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.744838 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.744846 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:48Z","lastTransitionTime":"2025-12-03T10:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.847095 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.847121 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.847130 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.847141 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.847149 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:48Z","lastTransitionTime":"2025-12-03T10:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.847178 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:48 crc kubenswrapper[4646]: E1203 10:54:48.847273 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.949513 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.949915 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.949926 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.949941 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:48 crc kubenswrapper[4646]: I1203 10:54:48.949952 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:48Z","lastTransitionTime":"2025-12-03T10:54:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.051746 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.051778 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.051789 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.051805 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.051817 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:49Z","lastTransitionTime":"2025-12-03T10:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.154132 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.154177 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.154193 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.154210 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.154221 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:49Z","lastTransitionTime":"2025-12-03T10:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.257065 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.257103 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.257113 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.257127 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.257145 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:49Z","lastTransitionTime":"2025-12-03T10:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.290441 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cbp2x_0b003d65-42ef-4a54-96e8-33428e114a1d/kube-multus/0.log" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.290489 4646 generic.go:334] "Generic (PLEG): container finished" podID="0b003d65-42ef-4a54-96e8-33428e114a1d" containerID="21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f" exitCode=1 Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.290514 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cbp2x" event={"ID":"0b003d65-42ef-4a54-96e8-33428e114a1d","Type":"ContainerDied","Data":"21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f"} Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.290847 4646 scope.go:117] "RemoveContainer" containerID="21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.304910 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.316864 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:48Z\\\",\\\"message\\\":\\\"2025-12-03T10:54:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_dd087b45-aa36-4e85-a4f7-be4971a083c8\\\\n2025-12-03T10:54:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_dd087b45-aa36-4e85-a4f7-be4971a083c8 to /host/opt/cni/bin/\\\\n2025-12-03T10:54:03Z [verbose] multus-daemon started\\\\n2025-12-03T10:54:03Z [verbose] Readiness Indicator file check\\\\n2025-12-03T10:54:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.327777 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.344453 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:31Z\\\",\\\"message\\\":\\\"o-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:31.771666 6174 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:31.772446 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 10:54:31.772522 6174 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:31.772536 6174 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:31.772544 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:31.772569 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:31.772589 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:31.772615 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:31.772621 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:31.772647 6174 factory.go:656] Stopping watch factory\\\\nI1203 10:54:31.772671 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 10:54:31.772680 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:31.772688 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:31.772703 6174 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.359252 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.359277 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.359285 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.359298 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.359309 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:49Z","lastTransitionTime":"2025-12-03T10:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.363778 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.378413 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.390660 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e90223ba-3577-408c-a674-92a9aaebc640\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8cbb08ab6f4da24ce8dcbafd255b9dd3461bc80abbd6d750e34abcbac41bb71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://821dd6864300c8286093c2ac960632ba03ed444f90994aa9cc24080e93472a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30f13d59d6c9c0046b62a4e9b202ba22e20014327ebd11dc59dc871a318826c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.405672 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.416203 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b171e4-cc29-4118-908c-d13825362e1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.426869 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.438915 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.449346 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.461180 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.461216 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.461230 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.461246 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.461258 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:49Z","lastTransitionTime":"2025-12-03T10:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.473918 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.484946 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.499167 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.519493 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.539453 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:49Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.563697 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.563729 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.563738 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.563751 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.563761 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:49Z","lastTransitionTime":"2025-12-03T10:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.666031 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.666070 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.666082 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.666133 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.666149 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:49Z","lastTransitionTime":"2025-12-03T10:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.774925 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.774972 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.774984 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.775001 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.775012 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:49Z","lastTransitionTime":"2025-12-03T10:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.847738 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.847792 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.847806 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:49 crc kubenswrapper[4646]: E1203 10:54:49.848016 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:49 crc kubenswrapper[4646]: E1203 10:54:49.848081 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:49 crc kubenswrapper[4646]: E1203 10:54:49.848173 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.877492 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.877520 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.877530 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.877542 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.877552 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:49Z","lastTransitionTime":"2025-12-03T10:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.979763 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.979807 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.979822 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.979838 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:49 crc kubenswrapper[4646]: I1203 10:54:49.979847 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:49Z","lastTransitionTime":"2025-12-03T10:54:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.081685 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.081737 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.081750 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.081768 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.081781 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:50Z","lastTransitionTime":"2025-12-03T10:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.184422 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.184462 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.184473 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.184490 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.184500 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:50Z","lastTransitionTime":"2025-12-03T10:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.286509 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.286571 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.286593 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.286617 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.286634 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:50Z","lastTransitionTime":"2025-12-03T10:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.294702 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cbp2x_0b003d65-42ef-4a54-96e8-33428e114a1d/kube-multus/0.log" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.294769 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cbp2x" event={"ID":"0b003d65-42ef-4a54-96e8-33428e114a1d","Type":"ContainerStarted","Data":"ffa715e7e498fda77dee5e1e92dece871ede3ada1b73f41f8f35ff15fc8910fd"} Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.306390 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.320225 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.331034 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b171e4-cc29-4118-908c-d13825362e1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.343736 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.357545 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.369148 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.380618 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.389504 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.389556 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.389574 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.389600 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.389617 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:50Z","lastTransitionTime":"2025-12-03T10:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.394285 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.405219 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.420782 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.433213 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.444402 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.458597 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e90223ba-3577-408c-a674-92a9aaebc640\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8cbb08ab6f4da24ce8dcbafd255b9dd3461bc80abbd6d750e34abcbac41bb71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://821dd6864300c8286093c2ac960632ba03ed444f90994aa9cc24080e93472a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30f13d59d6c9c0046b62a4e9b202ba22e20014327ebd11dc59dc871a318826c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.472581 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.483430 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffa715e7e498fda77dee5e1e92dece871ede3ada1b73f41f8f35ff15fc8910fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:48Z\\\",\\\"message\\\":\\\"2025-12-03T10:54:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_dd087b45-aa36-4e85-a4f7-be4971a083c8\\\\n2025-12-03T10:54:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_dd087b45-aa36-4e85-a4f7-be4971a083c8 to /host/opt/cni/bin/\\\\n2025-12-03T10:54:03Z [verbose] multus-daemon started\\\\n2025-12-03T10:54:03Z [verbose] Readiness Indicator file check\\\\n2025-12-03T10:54:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.491823 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.491857 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.491866 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.491881 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.491890 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:50Z","lastTransitionTime":"2025-12-03T10:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.494386 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.516070 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:31Z\\\",\\\"message\\\":\\\"o-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:31.771666 6174 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:31.772446 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 10:54:31.772522 6174 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:31.772536 6174 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:31.772544 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:31.772569 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:31.772589 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:31.772615 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:31.772621 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:31.772647 6174 factory.go:656] Stopping watch factory\\\\nI1203 10:54:31.772671 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 10:54:31.772680 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:31.772688 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:31.772703 6174 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:50Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.594554 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.594584 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.594593 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.594607 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.594621 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:50Z","lastTransitionTime":"2025-12-03T10:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.696847 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.696889 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.696898 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.696912 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.696921 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:50Z","lastTransitionTime":"2025-12-03T10:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.799961 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.800012 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.800021 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.800036 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.800047 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:50Z","lastTransitionTime":"2025-12-03T10:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.847849 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:50 crc kubenswrapper[4646]: E1203 10:54:50.847964 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.902169 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.902216 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.902226 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.902246 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:50 crc kubenswrapper[4646]: I1203 10:54:50.902258 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:50Z","lastTransitionTime":"2025-12-03T10:54:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.005504 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.005544 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.005557 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.005574 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.005587 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:51Z","lastTransitionTime":"2025-12-03T10:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.108215 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.108511 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.108594 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.108673 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.108740 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:51Z","lastTransitionTime":"2025-12-03T10:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.210994 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.211043 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.211053 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.211068 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.211079 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:51Z","lastTransitionTime":"2025-12-03T10:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.314137 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.314768 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.314832 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.314906 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.314982 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:51Z","lastTransitionTime":"2025-12-03T10:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.417405 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.417650 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.417715 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.417776 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.417841 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:51Z","lastTransitionTime":"2025-12-03T10:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.520469 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.520511 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.520523 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.520540 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.520552 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:51Z","lastTransitionTime":"2025-12-03T10:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.622665 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.622702 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.622710 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.622724 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.622734 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:51Z","lastTransitionTime":"2025-12-03T10:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.725195 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.725231 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.725241 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.725257 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.725266 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:51Z","lastTransitionTime":"2025-12-03T10:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.896749 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:51 crc kubenswrapper[4646]: E1203 10:54:51.896893 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.897477 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.897681 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:51 crc kubenswrapper[4646]: E1203 10:54:51.897763 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:51 crc kubenswrapper[4646]: E1203 10:54:51.897834 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.898389 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.898411 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.898421 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.898436 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:51 crc kubenswrapper[4646]: I1203 10:54:51.898445 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:51Z","lastTransitionTime":"2025-12-03T10:54:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.000261 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.000298 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.000313 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.000344 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.000354 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:52Z","lastTransitionTime":"2025-12-03T10:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.102075 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.102102 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.102110 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.102123 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.102131 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:52Z","lastTransitionTime":"2025-12-03T10:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.205669 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.205711 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.205724 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.205739 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.205751 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:52Z","lastTransitionTime":"2025-12-03T10:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.307578 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.307604 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.307612 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.307625 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.307634 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:52Z","lastTransitionTime":"2025-12-03T10:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.410092 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.410126 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.410138 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.410156 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.410167 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:52Z","lastTransitionTime":"2025-12-03T10:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.512922 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.512995 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.513008 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.513027 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.513043 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:52Z","lastTransitionTime":"2025-12-03T10:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.614995 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.615062 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.615074 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.615089 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.615100 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:52Z","lastTransitionTime":"2025-12-03T10:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.717247 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.717273 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.717281 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.717294 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.717303 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:52Z","lastTransitionTime":"2025-12-03T10:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.819374 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.819402 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.819437 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.819453 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.819462 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:52Z","lastTransitionTime":"2025-12-03T10:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.848053 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:52 crc kubenswrapper[4646]: E1203 10:54:52.848163 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.922026 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.922106 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.922121 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.922176 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:52 crc kubenswrapper[4646]: I1203 10:54:52.922194 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:52Z","lastTransitionTime":"2025-12-03T10:54:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.024717 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.024758 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.024769 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.024790 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.024802 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:53Z","lastTransitionTime":"2025-12-03T10:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.127215 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.127298 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.127316 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.127367 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.127392 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:53Z","lastTransitionTime":"2025-12-03T10:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.230228 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.230308 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.230326 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.230369 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.230388 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:53Z","lastTransitionTime":"2025-12-03T10:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.332051 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.332106 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.332125 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.332147 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.332166 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:53Z","lastTransitionTime":"2025-12-03T10:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.434858 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.434894 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.434904 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.434920 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.434931 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:53Z","lastTransitionTime":"2025-12-03T10:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.542382 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.542423 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.542431 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.542449 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.542459 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:53Z","lastTransitionTime":"2025-12-03T10:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.644683 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.644724 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.644738 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.644771 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.644787 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:53Z","lastTransitionTime":"2025-12-03T10:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.747684 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.747927 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.747987 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.748057 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.748114 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:53Z","lastTransitionTime":"2025-12-03T10:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.847397 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.847396 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:53 crc kubenswrapper[4646]: E1203 10:54:53.847531 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:53 crc kubenswrapper[4646]: E1203 10:54:53.847600 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.850028 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:53 crc kubenswrapper[4646]: E1203 10:54:53.850223 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.858815 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.859039 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.859102 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.859163 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.859215 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:53Z","lastTransitionTime":"2025-12-03T10:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.859549 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:53Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.870524 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e90223ba-3577-408c-a674-92a9aaebc640\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8cbb08ab6f4da24ce8dcbafd255b9dd3461bc80abbd6d750e34abcbac41bb71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://821dd6864300c8286093c2ac960632ba03ed444f90994aa9cc24080e93472a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30f13d59d6c9c0046b62a4e9b202ba22e20014327ebd11dc59dc871a318826c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:53Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.882610 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:53Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.897444 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffa715e7e498fda77dee5e1e92dece871ede3ada1b73f41f8f35ff15fc8910fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:48Z\\\",\\\"message\\\":\\\"2025-12-03T10:54:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_dd087b45-aa36-4e85-a4f7-be4971a083c8\\\\n2025-12-03T10:54:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_dd087b45-aa36-4e85-a4f7-be4971a083c8 to /host/opt/cni/bin/\\\\n2025-12-03T10:54:03Z [verbose] multus-daemon started\\\\n2025-12-03T10:54:03Z [verbose] Readiness Indicator file check\\\\n2025-12-03T10:54:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:53Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.912501 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:53Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.935584 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:31Z\\\",\\\"message\\\":\\\"o-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:31.771666 6174 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:31.772446 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 10:54:31.772522 6174 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:31.772536 6174 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:31.772544 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:31.772569 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:31.772589 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:31.772615 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:31.772621 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:31.772647 6174 factory.go:656] Stopping watch factory\\\\nI1203 10:54:31.772671 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 10:54:31.772680 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:31.772688 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:31.772703 6174 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:53Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.948173 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:53Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.961821 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:53Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.962017 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.962059 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.962072 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.962090 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.962103 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:53Z","lastTransitionTime":"2025-12-03T10:54:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.973273 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:53Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.982360 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b171e4-cc29-4118-908c-d13825362e1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:53Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:53 crc kubenswrapper[4646]: I1203 10:54:53.992810 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:53Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.002932 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:54Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.013065 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:54Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.021468 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:54Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.032257 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:54Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.040470 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:54Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.051589 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:54Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.063963 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.064021 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.064035 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.064052 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.064063 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:54Z","lastTransitionTime":"2025-12-03T10:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.167004 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.167046 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.167059 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.167075 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.167087 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:54Z","lastTransitionTime":"2025-12-03T10:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.269198 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.269244 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.269256 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.269274 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.269285 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:54Z","lastTransitionTime":"2025-12-03T10:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.371237 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.371278 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.371290 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.371306 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.371319 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:54Z","lastTransitionTime":"2025-12-03T10:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.475006 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.475041 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.475052 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.475071 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.475087 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:54Z","lastTransitionTime":"2025-12-03T10:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.577109 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.577154 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.577166 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.577183 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.577196 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:54Z","lastTransitionTime":"2025-12-03T10:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.679449 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.679481 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.679489 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.679503 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.679513 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:54Z","lastTransitionTime":"2025-12-03T10:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.781708 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.781764 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.781785 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.781814 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.781835 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:54Z","lastTransitionTime":"2025-12-03T10:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.847718 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:54 crc kubenswrapper[4646]: E1203 10:54:54.847924 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.884853 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.884911 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.884932 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.884957 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.884974 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:54Z","lastTransitionTime":"2025-12-03T10:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.987358 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.987618 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.987632 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.987647 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:54 crc kubenswrapper[4646]: I1203 10:54:54.987657 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:54Z","lastTransitionTime":"2025-12-03T10:54:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.089840 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.090042 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.090112 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.090195 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.090257 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:55Z","lastTransitionTime":"2025-12-03T10:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.192018 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.192075 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.192087 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.192102 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.192112 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:55Z","lastTransitionTime":"2025-12-03T10:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.294701 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.294911 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.294993 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.295077 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.295167 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:55Z","lastTransitionTime":"2025-12-03T10:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.397657 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.397937 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.398054 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.398137 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.398207 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:55Z","lastTransitionTime":"2025-12-03T10:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.502464 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.503116 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.503324 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.503519 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.503667 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:55Z","lastTransitionTime":"2025-12-03T10:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.605911 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.606144 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.606209 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.606275 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.606351 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:55Z","lastTransitionTime":"2025-12-03T10:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.708776 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.708819 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.708830 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.708848 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.708860 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:55Z","lastTransitionTime":"2025-12-03T10:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.812164 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.812211 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.812227 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.812246 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.812258 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:55Z","lastTransitionTime":"2025-12-03T10:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.853172 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:55 crc kubenswrapper[4646]: E1203 10:54:55.853387 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.853710 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:55 crc kubenswrapper[4646]: E1203 10:54:55.854020 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.854314 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:55 crc kubenswrapper[4646]: E1203 10:54:55.854473 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.874637 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.874674 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.874691 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.874712 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.874730 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:55Z","lastTransitionTime":"2025-12-03T10:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:55 crc kubenswrapper[4646]: E1203 10:54:55.892398 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:55Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.898109 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.898193 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.898212 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.898270 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.898295 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:55Z","lastTransitionTime":"2025-12-03T10:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:55 crc kubenswrapper[4646]: E1203 10:54:55.919794 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:55Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.924137 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.924243 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.924316 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.924405 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.924475 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:55Z","lastTransitionTime":"2025-12-03T10:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:55 crc kubenswrapper[4646]: E1203 10:54:55.941286 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:55Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.945942 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.945999 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.946018 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.946042 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.946096 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:55Z","lastTransitionTime":"2025-12-03T10:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:55 crc kubenswrapper[4646]: E1203 10:54:55.965241 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:55Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.969886 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.969989 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.970062 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.970161 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.970220 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:55Z","lastTransitionTime":"2025-12-03T10:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:55 crc kubenswrapper[4646]: E1203 10:54:55.988057 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:55Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:55 crc kubenswrapper[4646]: E1203 10:54:55.988206 4646 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.989803 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.989833 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.989845 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.989861 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:55 crc kubenswrapper[4646]: I1203 10:54:55.989873 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:55Z","lastTransitionTime":"2025-12-03T10:54:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.092428 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.092455 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.092463 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.092475 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.092485 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:56Z","lastTransitionTime":"2025-12-03T10:54:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.194990 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.195254 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.195371 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.195491 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.195574 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:56Z","lastTransitionTime":"2025-12-03T10:54:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.298605 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.298651 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.298664 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.298684 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.298698 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:56Z","lastTransitionTime":"2025-12-03T10:54:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.401013 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.401294 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.401406 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.401494 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.401623 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:56Z","lastTransitionTime":"2025-12-03T10:54:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.503742 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.503805 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.503816 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.503840 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.503850 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:56Z","lastTransitionTime":"2025-12-03T10:54:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.606518 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.606822 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.606932 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.607020 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.607091 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:56Z","lastTransitionTime":"2025-12-03T10:54:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.710122 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.710452 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.710535 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.710609 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.710699 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:56Z","lastTransitionTime":"2025-12-03T10:54:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.813820 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.813890 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.813907 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.813933 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.813950 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:56Z","lastTransitionTime":"2025-12-03T10:54:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.847669 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:56 crc kubenswrapper[4646]: E1203 10:54:56.847993 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.916472 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.916767 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.916904 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.917222 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:56 crc kubenswrapper[4646]: I1203 10:54:56.917406 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:56Z","lastTransitionTime":"2025-12-03T10:54:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.020809 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.020873 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.020891 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.020916 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.020931 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:57Z","lastTransitionTime":"2025-12-03T10:54:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.123300 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.123689 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.123840 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.123987 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.124128 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:57Z","lastTransitionTime":"2025-12-03T10:54:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.226822 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.226854 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.226866 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.226883 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.226895 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:57Z","lastTransitionTime":"2025-12-03T10:54:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.329387 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.329438 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.329456 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.329478 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.329495 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:57Z","lastTransitionTime":"2025-12-03T10:54:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.432532 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.432566 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.432576 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.432594 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.432607 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:57Z","lastTransitionTime":"2025-12-03T10:54:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.535361 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.535431 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.535446 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.535467 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.535481 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:57Z","lastTransitionTime":"2025-12-03T10:54:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.638137 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.638191 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.638208 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.638233 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.638250 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:57Z","lastTransitionTime":"2025-12-03T10:54:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.741305 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.741462 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.741544 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.741678 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.741843 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:57Z","lastTransitionTime":"2025-12-03T10:54:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.844600 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.844663 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.844691 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.844725 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.844748 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:57Z","lastTransitionTime":"2025-12-03T10:54:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.848225 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:57 crc kubenswrapper[4646]: E1203 10:54:57.848459 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.848478 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.848785 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:57 crc kubenswrapper[4646]: E1203 10:54:57.848965 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:57 crc kubenswrapper[4646]: E1203 10:54:57.849611 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.850387 4646 scope.go:117] "RemoveContainer" containerID="bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.948120 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.948633 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.948656 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.948682 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:57 crc kubenswrapper[4646]: I1203 10:54:57.948705 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:57Z","lastTransitionTime":"2025-12-03T10:54:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.051817 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.051893 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.051914 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.051936 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.051975 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:58Z","lastTransitionTime":"2025-12-03T10:54:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.155617 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.155665 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.155679 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.155698 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.155712 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:58Z","lastTransitionTime":"2025-12-03T10:54:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.258096 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.258159 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.258176 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.258198 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.258214 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:58Z","lastTransitionTime":"2025-12-03T10:54:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.324273 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovnkube-controller/2.log" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.327212 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerStarted","Data":"8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034"} Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.328247 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.344068 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.361211 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.361238 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.361245 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.361258 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.361267 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:58Z","lastTransitionTime":"2025-12-03T10:54:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.362864 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.387539 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.400889 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e90223ba-3577-408c-a674-92a9aaebc640\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8cbb08ab6f4da24ce8dcbafd255b9dd3461bc80abbd6d750e34abcbac41bb71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://821dd6864300c8286093c2ac960632ba03ed444f90994aa9cc24080e93472a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30f13d59d6c9c0046b62a4e9b202ba22e20014327ebd11dc59dc871a318826c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.424900 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.438499 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffa715e7e498fda77dee5e1e92dece871ede3ada1b73f41f8f35ff15fc8910fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:48Z\\\",\\\"message\\\":\\\"2025-12-03T10:54:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_dd087b45-aa36-4e85-a4f7-be4971a083c8\\\\n2025-12-03T10:54:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_dd087b45-aa36-4e85-a4f7-be4971a083c8 to /host/opt/cni/bin/\\\\n2025-12-03T10:54:03Z [verbose] multus-daemon started\\\\n2025-12-03T10:54:03Z [verbose] Readiness Indicator file check\\\\n2025-12-03T10:54:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.453571 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.462828 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.462858 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.462868 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.462883 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.462893 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:58Z","lastTransitionTime":"2025-12-03T10:54:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.476254 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:31Z\\\",\\\"message\\\":\\\"o-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:31.771666 6174 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:31.772446 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 10:54:31.772522 6174 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:31.772536 6174 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:31.772544 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:31.772569 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:31.772589 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:31.772615 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:31.772621 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:31.772647 6174 factory.go:656] Stopping watch factory\\\\nI1203 10:54:31.772671 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 10:54:31.772680 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:31.772688 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:31.772703 6174 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.487110 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.500361 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.530664 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b171e4-cc29-4118-908c-d13825362e1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.548991 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.560376 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.564635 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.564670 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.564681 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.564696 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.564707 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:58Z","lastTransitionTime":"2025-12-03T10:54:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.575104 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.583417 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.596463 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.605724 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.666562 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.666597 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.666605 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.666619 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.666628 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:58Z","lastTransitionTime":"2025-12-03T10:54:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.768892 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.768922 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.768930 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.768944 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.768954 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:58Z","lastTransitionTime":"2025-12-03T10:54:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.847634 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:54:58 crc kubenswrapper[4646]: E1203 10:54:58.847781 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.871722 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.871783 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.871800 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.871825 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.871842 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:58Z","lastTransitionTime":"2025-12-03T10:54:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.975149 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.975177 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.975184 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.975197 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:58 crc kubenswrapper[4646]: I1203 10:54:58.975206 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:58Z","lastTransitionTime":"2025-12-03T10:54:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.078415 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.078482 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.078500 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.078527 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.078546 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:59Z","lastTransitionTime":"2025-12-03T10:54:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.180853 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.180891 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.180902 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.180918 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.180929 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:59Z","lastTransitionTime":"2025-12-03T10:54:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.284086 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.284121 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.284129 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.284148 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.284158 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:59Z","lastTransitionTime":"2025-12-03T10:54:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.332851 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovnkube-controller/3.log" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.333643 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovnkube-controller/2.log" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.336841 4646 generic.go:334] "Generic (PLEG): container finished" podID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerID="8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034" exitCode=1 Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.336895 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerDied","Data":"8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034"} Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.336951 4646 scope.go:117] "RemoveContainer" containerID="bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.338185 4646 scope.go:117] "RemoveContainer" containerID="8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034" Dec 03 10:54:59 crc kubenswrapper[4646]: E1203 10:54:59.338527 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.363039 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b171e4-cc29-4118-908c-d13825362e1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.382218 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.387839 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.387872 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.387881 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.387896 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.387908 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:59Z","lastTransitionTime":"2025-12-03T10:54:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.398362 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.411241 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.428641 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.444112 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.458494 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.476431 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.489260 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.490092 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.490119 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.490127 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.490140 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.490150 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:59Z","lastTransitionTime":"2025-12-03T10:54:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.505289 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.520823 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffa715e7e498fda77dee5e1e92dece871ede3ada1b73f41f8f35ff15fc8910fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:48Z\\\",\\\"message\\\":\\\"2025-12-03T10:54:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_dd087b45-aa36-4e85-a4f7-be4971a083c8\\\\n2025-12-03T10:54:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_dd087b45-aa36-4e85-a4f7-be4971a083c8 to /host/opt/cni/bin/\\\\n2025-12-03T10:54:03Z [verbose] multus-daemon started\\\\n2025-12-03T10:54:03Z [verbose] Readiness Indicator file check\\\\n2025-12-03T10:54:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.533215 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.549152 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd6f454e2162626dedbc53be57d4002007d5aaa4a762311d78e1dd43f82260a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:31Z\\\",\\\"message\\\":\\\"o-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 10:54:31.771666 6174 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 10:54:31.772446 6174 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 10:54:31.772522 6174 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 10:54:31.772536 6174 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 10:54:31.772544 6174 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 10:54:31.772569 6174 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 10:54:31.772589 6174 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 10:54:31.772615 6174 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 10:54:31.772621 6174 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 10:54:31.772647 6174 factory.go:656] Stopping watch factory\\\\nI1203 10:54:31.772671 6174 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 10:54:31.772680 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 10:54:31.772688 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 10:54:31.772703 6174 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:58Z\\\",\\\"message\\\":\\\"d for openshift-multus/network-metrics-daemon-rs5rq: failed to update pod openshift-multus/network-metrics-daemon-rs5rq: Internal error occurred: failed calling webhook \\\\\\\"pod.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/pod?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z\\\\nI1203 10:54:58.900630 6513 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 2.826719ms\\\\nF1203 10:54:58.900626 6513 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certifi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.566115 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.581234 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.592436 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.592473 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.592484 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.592501 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.592512 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:59Z","lastTransitionTime":"2025-12-03T10:54:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.596234 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e90223ba-3577-408c-a674-92a9aaebc640\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8cbb08ab6f4da24ce8dcbafd255b9dd3461bc80abbd6d750e34abcbac41bb71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://821dd6864300c8286093c2ac960632ba03ed444f90994aa9cc24080e93472a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30f13d59d6c9c0046b62a4e9b202ba22e20014327ebd11dc59dc871a318826c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.609444 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:59Z is after 2025-08-24T17:21:41Z" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.694972 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.695012 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.695046 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.695065 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.695075 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:59Z","lastTransitionTime":"2025-12-03T10:54:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.799071 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.799108 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.799119 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.799137 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.799147 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:59Z","lastTransitionTime":"2025-12-03T10:54:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.847789 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:54:59 crc kubenswrapper[4646]: E1203 10:54:59.847985 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.848278 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:54:59 crc kubenswrapper[4646]: E1203 10:54:59.848413 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.848711 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:54:59 crc kubenswrapper[4646]: E1203 10:54:59.848931 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.902848 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.902938 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.902957 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.903165 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:54:59 crc kubenswrapper[4646]: I1203 10:54:59.903185 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:54:59Z","lastTransitionTime":"2025-12-03T10:54:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.006171 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.006247 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.006269 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.006300 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.006319 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:00Z","lastTransitionTime":"2025-12-03T10:55:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.109578 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.109685 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.109759 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.109795 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.109878 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:00Z","lastTransitionTime":"2025-12-03T10:55:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.212593 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.212626 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.212637 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.212653 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.212664 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:00Z","lastTransitionTime":"2025-12-03T10:55:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.315233 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.315268 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.315282 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.315301 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.315315 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:00Z","lastTransitionTime":"2025-12-03T10:55:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.342048 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovnkube-controller/3.log" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.347186 4646 scope.go:117] "RemoveContainer" containerID="8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034" Dec 03 10:55:00 crc kubenswrapper[4646]: E1203 10:55:00.347472 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.365500 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.386725 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.408886 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e90223ba-3577-408c-a674-92a9aaebc640\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8cbb08ab6f4da24ce8dcbafd255b9dd3461bc80abbd6d750e34abcbac41bb71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://821dd6864300c8286093c2ac960632ba03ed444f90994aa9cc24080e93472a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30f13d59d6c9c0046b62a4e9b202ba22e20014327ebd11dc59dc871a318826c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.418139 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.418203 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.418222 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.418257 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.418274 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:00Z","lastTransitionTime":"2025-12-03T10:55:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.424682 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.441542 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffa715e7e498fda77dee5e1e92dece871ede3ada1b73f41f8f35ff15fc8910fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:48Z\\\",\\\"message\\\":\\\"2025-12-03T10:54:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_dd087b45-aa36-4e85-a4f7-be4971a083c8\\\\n2025-12-03T10:54:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_dd087b45-aa36-4e85-a4f7-be4971a083c8 to /host/opt/cni/bin/\\\\n2025-12-03T10:54:03Z [verbose] multus-daemon started\\\\n2025-12-03T10:54:03Z [verbose] Readiness Indicator file check\\\\n2025-12-03T10:54:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.457061 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.479546 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:58Z\\\",\\\"message\\\":\\\"d for openshift-multus/network-metrics-daemon-rs5rq: failed to update pod openshift-multus/network-metrics-daemon-rs5rq: Internal error occurred: failed calling webhook \\\\\\\"pod.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/pod?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z\\\\nI1203 10:54:58.900630 6513 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 2.826719ms\\\\nF1203 10:54:58.900626 6513 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certifi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.497842 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.509119 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.521027 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.521085 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.521102 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.521124 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.521140 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:00Z","lastTransitionTime":"2025-12-03T10:55:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.523196 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.536327 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b171e4-cc29-4118-908c-d13825362e1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.548883 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.561367 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.572114 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.580769 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.589924 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.601081 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:00Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.623607 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.623842 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.623951 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.624041 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.624146 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:00Z","lastTransitionTime":"2025-12-03T10:55:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.726953 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.726986 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.726996 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.727010 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.727019 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:00Z","lastTransitionTime":"2025-12-03T10:55:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.830379 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.830472 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.830492 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.830519 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.830537 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:00Z","lastTransitionTime":"2025-12-03T10:55:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.847574 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:00 crc kubenswrapper[4646]: E1203 10:55:00.847958 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.933553 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.933612 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.933632 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.933656 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:00 crc kubenswrapper[4646]: I1203 10:55:00.933673 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:00Z","lastTransitionTime":"2025-12-03T10:55:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.037520 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.037582 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.037598 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.037622 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.037642 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:01Z","lastTransitionTime":"2025-12-03T10:55:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.141378 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.141435 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.141455 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.141479 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.141499 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:01Z","lastTransitionTime":"2025-12-03T10:55:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.244166 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.244199 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.244209 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.244224 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.244235 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:01Z","lastTransitionTime":"2025-12-03T10:55:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.352221 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.352297 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.352319 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.352372 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.352400 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:01Z","lastTransitionTime":"2025-12-03T10:55:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.456414 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.456814 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.457020 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.457201 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.457416 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:01Z","lastTransitionTime":"2025-12-03T10:55:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.561141 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.561566 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.562133 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.562199 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.562221 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:01Z","lastTransitionTime":"2025-12-03T10:55:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.664942 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.664985 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.664996 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.665011 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.665022 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:01Z","lastTransitionTime":"2025-12-03T10:55:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.767604 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.767638 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.767647 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.767669 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.767682 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:01Z","lastTransitionTime":"2025-12-03T10:55:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.847588 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.847686 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.847690 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:01 crc kubenswrapper[4646]: E1203 10:55:01.847821 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:01 crc kubenswrapper[4646]: E1203 10:55:01.848178 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:01 crc kubenswrapper[4646]: E1203 10:55:01.848498 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.870456 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.870517 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.870540 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.870566 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.870588 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:01Z","lastTransitionTime":"2025-12-03T10:55:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.973535 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.973614 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.973638 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.973665 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:01 crc kubenswrapper[4646]: I1203 10:55:01.973686 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:01Z","lastTransitionTime":"2025-12-03T10:55:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.076568 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.076686 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.076710 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.076741 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.076762 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:02Z","lastTransitionTime":"2025-12-03T10:55:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.179749 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.179827 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.179847 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.179873 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.179896 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:02Z","lastTransitionTime":"2025-12-03T10:55:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.282710 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.282795 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.282826 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.282855 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.282878 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:02Z","lastTransitionTime":"2025-12-03T10:55:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.385002 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.385079 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.385099 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.385148 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.385166 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:02Z","lastTransitionTime":"2025-12-03T10:55:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.488371 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.488424 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.488441 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.488466 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.488485 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:02Z","lastTransitionTime":"2025-12-03T10:55:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.591577 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.591627 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.591644 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.591668 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.591685 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:02Z","lastTransitionTime":"2025-12-03T10:55:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.695523 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.695587 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.695611 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.695645 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.695668 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:02Z","lastTransitionTime":"2025-12-03T10:55:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.798671 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.798731 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.798757 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.798782 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.798798 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:02Z","lastTransitionTime":"2025-12-03T10:55:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.847397 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:02 crc kubenswrapper[4646]: E1203 10:55:02.847578 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.902285 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.902361 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.902379 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.902405 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:02 crc kubenswrapper[4646]: I1203 10:55:02.902423 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:02Z","lastTransitionTime":"2025-12-03T10:55:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.004933 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.004992 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.005013 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.005039 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.005055 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:03Z","lastTransitionTime":"2025-12-03T10:55:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.107730 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.107796 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.107811 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.107829 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.107861 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:03Z","lastTransitionTime":"2025-12-03T10:55:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.210998 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.211059 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.211078 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.211103 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.211120 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:03Z","lastTransitionTime":"2025-12-03T10:55:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.314317 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.314642 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.314784 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.314940 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.315062 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:03Z","lastTransitionTime":"2025-12-03T10:55:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.418461 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.418526 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.418545 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.418571 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.418590 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:03Z","lastTransitionTime":"2025-12-03T10:55:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.524822 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.525205 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.525723 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.526132 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.526393 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:03Z","lastTransitionTime":"2025-12-03T10:55:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.629172 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.629240 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.629261 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.629331 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.629386 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:03Z","lastTransitionTime":"2025-12-03T10:55:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.731492 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.731785 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.731944 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.732107 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.732242 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:03Z","lastTransitionTime":"2025-12-03T10:55:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.834830 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.834898 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.834920 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.834947 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.834964 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:03Z","lastTransitionTime":"2025-12-03T10:55:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.848173 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.848200 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:03 crc kubenswrapper[4646]: E1203 10:55:03.848402 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.848482 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:03 crc kubenswrapper[4646]: E1203 10:55:03.848589 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:03 crc kubenswrapper[4646]: E1203 10:55:03.848687 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.872170 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8869765-e63a-44d3-912f-8db6a978db59\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"message\\\":\\\"le observer\\\\nW1203 10:54:01.216809 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1203 10:54:01.216897 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 10:54:01.218608 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3362096582/tls.crt::/tmp/serving-cert-3362096582/tls.key\\\\\\\"\\\\nI1203 10:54:01.990360 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 10:54:01.993125 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 10:54:01.993144 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 10:54:01.993172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 10:54:01.993193 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 10:54:01.997642 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 10:54:01.997661 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997666 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 10:54:01.997670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 10:54:01.997673 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 10:54:01.997676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 10:54:01.997679 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 10:54:01.997842 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 10:54:01.998750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.893025 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.912266 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cbp2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b003d65-42ef-4a54-96e8-33428e114a1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ffa715e7e498fda77dee5e1e92dece871ede3ada1b73f41f8f35ff15fc8910fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:48Z\\\",\\\"message\\\":\\\"2025-12-03T10:54:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_dd087b45-aa36-4e85-a4f7-be4971a083c8\\\\n2025-12-03T10:54:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_dd087b45-aa36-4e85-a4f7-be4971a083c8 to /host/opt/cni/bin/\\\\n2025-12-03T10:54:03Z [verbose] multus-daemon started\\\\n2025-12-03T10:54:03Z [verbose] Readiness Indicator file check\\\\n2025-12-03T10:54:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cbp2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.929708 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9695d947-fc43-4625-86af-a1a3e177bb27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bc8a8808774c579f262b564e8721cda04d0bf00b6eed8b059ba4f88ab27edcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6gg9w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pndzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.936717 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.936768 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.936778 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.936796 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.936808 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:03Z","lastTransitionTime":"2025-12-03T10:55:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.951261 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cd1981e-44c4-4251-81d1-c07c9d45258b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T10:54:58Z\\\",\\\"message\\\":\\\"d for openshift-multus/network-metrics-daemon-rs5rq: failed to update pod openshift-multus/network-metrics-daemon-rs5rq: Internal error occurred: failed calling webhook \\\\\\\"pod.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/pod?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:54:58Z is after 2025-08-24T17:21:41Z\\\\nI1203 10:54:58.900630 6513 services_controller.go:360] Finished syncing service olm-operator-metrics on namespace openshift-operator-lifecycle-manager for network=default : 2.826719ms\\\\nF1203 10:54:58.900626 6513 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certifi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jct4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9lfpq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.972920 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b49xq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fef52874-4a7c-4016-938a-69e88790ab19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39b579eca2eee912ac9dfeda82816fda28f3bca58214d9559d906ec0fa1e6504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09bd26272800fc58248a48b578a1a15301d7ac4196aface1713dfdab31424500\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d4d93495e43adb61b7354bad86e74a0488b509a92cb6bfa995a1247abf9f3a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5abd12e42a458fde473d835bde4cbb1faf57b273c00a9637af16891d239c22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://268861434adecd0ca4add048d88df95b35ebb41a22137bf7213ee70deadbdb8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9234bdfe8bafb6883e54bc6239d0f74e3e2d41a104a1e9cd13ff8b95ebf54817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33e3372105dd4352ee51410709e087bb2ca3536344153938d54cc0929e948be9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:54:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:54:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h29bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b49xq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.990261 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"552f7411-b8c5-4102-98eb-045286f05b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://549487b3700ad81332b89fd65d592f1bae0bc7f42d84732910af2e4487f390d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b719496e59b9434d06e81b4901df105d5422a9480e6c886fcf4bddd3ca37b53f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ceb8ca03c1111764f973f5f5024626733d2f8b1463015a006b376554621aa4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:03 crc kubenswrapper[4646]: I1203 10:55:03.999509 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e90223ba-3577-408c-a674-92a9aaebc640\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:53:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8cbb08ab6f4da24ce8dcbafd255b9dd3461bc80abbd6d750e34abcbac41bb71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://821dd6864300c8286093c2ac960632ba03ed444f90994aa9cc24080e93472a5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30f13d59d6c9c0046b62a4e9b202ba22e20014327ebd11dc59dc871a318826c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:53:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b7983be855761dcbca521f882404799eadf1eb411b8023af3ade205ca4d4a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T10:53:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T10:53:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:53:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:03Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.008223 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"263a8752-97c7-4d08-9822-1e85687116fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d63da7378b6dacdfec933f4203f4cf4438702ab4207466204d98b6eb331671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://143fae65cb2d9a01816b20c48e442931e3f268c92fba39ed60b450015ca0ac6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rhdl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gprb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.017417 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b171e4-cc29-4118-908c-d13825362e1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8kkt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs5rq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.028243 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4efd35b8491568e2ec78620846bb180cc83e375b28478250ecd9f734591f6a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.039202 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.039235 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.039245 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.039261 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.039263 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc4c622a0d747d721bc330526f9fa7e540d92c9b8b2e612f2d4aaf9b0ab3014d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.039271 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:04Z","lastTransitionTime":"2025-12-03T10:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.047029 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffm54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82e5903b-9412-4fe7-b248-4ad185229bbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06a709c4b1e55e0ebda2b3be4d79d38e5cf87a85785f827844d3a24f76a767ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4pgc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffm54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.056730 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3b7545ca6f94a7ec017e4a073d6b72901bb1f9d8467151c07f4ccb60dcd2098\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60d342e8f9c83171e94e7ec5ca67a14880d6c6239e4ec5c4cc383ca85e549929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.065122 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lhrsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4dc8da-e71d-4968-a2d4-1b03a616bac4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2cf1c669f85f7885e998b0945735af320de511f225ca79cb9e3eeaac031ea05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T10:54:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tjp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T10:54:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lhrsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.076895 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.089762 4646 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T10:54:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:04Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.142092 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.142114 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.142123 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.142136 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.142146 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:04Z","lastTransitionTime":"2025-12-03T10:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.244751 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.244837 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.244856 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.244917 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.244937 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:04Z","lastTransitionTime":"2025-12-03T10:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.347522 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.347546 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.347555 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.347582 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.347593 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:04Z","lastTransitionTime":"2025-12-03T10:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.450314 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.450408 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.450445 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.450474 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.450493 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:04Z","lastTransitionTime":"2025-12-03T10:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.553705 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.553767 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.553790 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.553821 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.553842 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:04Z","lastTransitionTime":"2025-12-03T10:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.657132 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.657193 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.657215 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.657244 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.657264 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:04Z","lastTransitionTime":"2025-12-03T10:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.760497 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.760551 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.760570 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.760596 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.760621 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:04Z","lastTransitionTime":"2025-12-03T10:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.847808 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:04 crc kubenswrapper[4646]: E1203 10:55:04.848019 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.863445 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.863737 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.863763 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.863781 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.863801 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.863818 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:04Z","lastTransitionTime":"2025-12-03T10:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.968032 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.968090 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.968106 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.968129 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:04 crc kubenswrapper[4646]: I1203 10:55:04.968145 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:04Z","lastTransitionTime":"2025-12-03T10:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.072088 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.072452 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.072670 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.072867 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.073080 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:05Z","lastTransitionTime":"2025-12-03T10:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.176125 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.176225 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.176662 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.176717 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.176737 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:05Z","lastTransitionTime":"2025-12-03T10:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.280608 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.280680 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.280703 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.280731 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.280750 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:05Z","lastTransitionTime":"2025-12-03T10:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.384233 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.384634 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.384876 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.385111 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.385375 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:05Z","lastTransitionTime":"2025-12-03T10:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.489058 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.489458 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.489628 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.489763 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.489952 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:05Z","lastTransitionTime":"2025-12-03T10:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.592759 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.592797 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.592811 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.592829 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.592842 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:05Z","lastTransitionTime":"2025-12-03T10:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.696246 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.696287 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.696303 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.696323 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.696354 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:05Z","lastTransitionTime":"2025-12-03T10:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.732979 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.733087 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.733128 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.733182 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.733206 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.733359 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.733380 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.733392 4646 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.733436 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 10:56:09.733420213 +0000 UTC m=+146.196476358 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.733498 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:09.733489505 +0000 UTC m=+146.196545650 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.733552 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.733562 4646 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.733571 4646 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.733597 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 10:56:09.733589158 +0000 UTC m=+146.196645303 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.733649 4646 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.733675 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:56:09.73366763 +0000 UTC m=+146.196723775 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.733709 4646 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.733732 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 10:56:09.733725262 +0000 UTC m=+146.196781407 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.806122 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.806192 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.806213 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.806280 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.806307 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:05Z","lastTransitionTime":"2025-12-03T10:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.848039 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.848140 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.848251 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.848209 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.848501 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:05 crc kubenswrapper[4646]: E1203 10:55:05.848677 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.910884 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.910920 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.910933 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.910959 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:05 crc kubenswrapper[4646]: I1203 10:55:05.910971 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:05Z","lastTransitionTime":"2025-12-03T10:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.014994 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.015360 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.015687 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.015820 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.015933 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:06Z","lastTransitionTime":"2025-12-03T10:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.118504 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.118571 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.118590 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.118616 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.118634 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:06Z","lastTransitionTime":"2025-12-03T10:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.222518 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.222600 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.222628 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.222663 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.222687 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:06Z","lastTransitionTime":"2025-12-03T10:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.226597 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.226637 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.226652 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.226671 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.226686 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:06Z","lastTransitionTime":"2025-12-03T10:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:06 crc kubenswrapper[4646]: E1203 10:55:06.247387 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.253174 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.253229 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.253246 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.253270 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.253288 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:06Z","lastTransitionTime":"2025-12-03T10:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:06 crc kubenswrapper[4646]: E1203 10:55:06.278287 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.284623 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.284735 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.284760 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.284794 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.284824 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:06Z","lastTransitionTime":"2025-12-03T10:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:06 crc kubenswrapper[4646]: E1203 10:55:06.308945 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.314781 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.315129 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.315156 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.315187 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.315210 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:06Z","lastTransitionTime":"2025-12-03T10:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:06 crc kubenswrapper[4646]: E1203 10:55:06.340594 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.346627 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.346685 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.346704 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.346729 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.346747 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:06Z","lastTransitionTime":"2025-12-03T10:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:06 crc kubenswrapper[4646]: E1203 10:55:06.372046 4646 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T10:55:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"555f41f0-f8e0-41bc-91f8-8b81b382648c\\\",\\\"systemUUID\\\":\\\"10008c78-4919-4934-8d65-42e9210ff073\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T10:55:06Z is after 2025-08-24T17:21:41Z" Dec 03 10:55:06 crc kubenswrapper[4646]: E1203 10:55:06.372280 4646 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.375212 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.375280 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.375305 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.375370 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.375406 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:06Z","lastTransitionTime":"2025-12-03T10:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.479196 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.479257 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.479275 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.479300 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.479387 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:06Z","lastTransitionTime":"2025-12-03T10:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.582955 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.583050 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.583069 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.583092 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.583109 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:06Z","lastTransitionTime":"2025-12-03T10:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.686463 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.686511 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.686531 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.686557 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.686573 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:06Z","lastTransitionTime":"2025-12-03T10:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.790057 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.790111 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.790129 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.790154 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.790172 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:06Z","lastTransitionTime":"2025-12-03T10:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.848003 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:06 crc kubenswrapper[4646]: E1203 10:55:06.848150 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.893381 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.893441 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.893457 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.893477 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.893493 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:06Z","lastTransitionTime":"2025-12-03T10:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.996038 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.996128 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.996147 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.996172 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:06 crc kubenswrapper[4646]: I1203 10:55:06.996189 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:06Z","lastTransitionTime":"2025-12-03T10:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.099220 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.099281 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.099302 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.099378 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.099409 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:07Z","lastTransitionTime":"2025-12-03T10:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.202311 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.202428 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.202452 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.202482 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.202504 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:07Z","lastTransitionTime":"2025-12-03T10:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.305594 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.305671 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.305690 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.305718 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.305738 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:07Z","lastTransitionTime":"2025-12-03T10:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.408286 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.408378 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.408408 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.408437 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.408454 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:07Z","lastTransitionTime":"2025-12-03T10:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.512464 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.512514 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.512531 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.512553 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.512573 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:07Z","lastTransitionTime":"2025-12-03T10:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.615677 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.615719 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.615730 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.615746 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.615758 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:07Z","lastTransitionTime":"2025-12-03T10:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.719029 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.719280 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.719358 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.719454 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.719519 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:07Z","lastTransitionTime":"2025-12-03T10:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.823158 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.824008 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.824166 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.824317 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.824513 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:07Z","lastTransitionTime":"2025-12-03T10:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.847770 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.847854 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.847991 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:07 crc kubenswrapper[4646]: E1203 10:55:07.847999 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:07 crc kubenswrapper[4646]: E1203 10:55:07.848128 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:07 crc kubenswrapper[4646]: E1203 10:55:07.848235 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.927198 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.927460 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.927540 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.927614 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:07 crc kubenswrapper[4646]: I1203 10:55:07.927681 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:07Z","lastTransitionTime":"2025-12-03T10:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.031202 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.031291 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.031308 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.031355 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.031375 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:08Z","lastTransitionTime":"2025-12-03T10:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.133835 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.133870 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.133880 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.133896 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.133908 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:08Z","lastTransitionTime":"2025-12-03T10:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.237556 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.237629 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.237650 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.237677 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.237694 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:08Z","lastTransitionTime":"2025-12-03T10:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.341670 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.341782 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.341801 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.341840 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.341876 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:08Z","lastTransitionTime":"2025-12-03T10:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.444921 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.444973 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.444992 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.445018 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.445034 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:08Z","lastTransitionTime":"2025-12-03T10:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.548198 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.548758 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.549018 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.549428 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.549896 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:08Z","lastTransitionTime":"2025-12-03T10:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.653924 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.654241 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.654462 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.654687 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.654813 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:08Z","lastTransitionTime":"2025-12-03T10:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.758802 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.759138 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.759313 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.759548 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.759717 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:08Z","lastTransitionTime":"2025-12-03T10:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.848121 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:08 crc kubenswrapper[4646]: E1203 10:55:08.848608 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.861863 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.861901 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.861914 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.861931 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.861942 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:08Z","lastTransitionTime":"2025-12-03T10:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.965004 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.965461 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.965714 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.965924 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:08 crc kubenswrapper[4646]: I1203 10:55:08.966402 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:08Z","lastTransitionTime":"2025-12-03T10:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.070048 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.070089 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.070099 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.070114 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.070123 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:09Z","lastTransitionTime":"2025-12-03T10:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.172461 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.172837 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.172967 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.173105 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.173232 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:09Z","lastTransitionTime":"2025-12-03T10:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.298715 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.299105 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.299305 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.299585 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.299868 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:09Z","lastTransitionTime":"2025-12-03T10:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.404762 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.405200 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.405620 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.405835 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.405995 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:09Z","lastTransitionTime":"2025-12-03T10:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.514163 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.514513 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.514604 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.514674 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.514796 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:09Z","lastTransitionTime":"2025-12-03T10:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.618287 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.618382 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.618405 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.618434 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.618455 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:09Z","lastTransitionTime":"2025-12-03T10:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.721529 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.721891 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.722058 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.722216 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.722501 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:09Z","lastTransitionTime":"2025-12-03T10:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.825654 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.825929 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.826071 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.826212 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.826367 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:09Z","lastTransitionTime":"2025-12-03T10:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.848327 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:09 crc kubenswrapper[4646]: E1203 10:55:09.848554 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.848830 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:09 crc kubenswrapper[4646]: E1203 10:55:09.848925 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.849645 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:09 crc kubenswrapper[4646]: E1203 10:55:09.849912 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.929620 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.929678 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.929695 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.929718 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:09 crc kubenswrapper[4646]: I1203 10:55:09.929734 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:09Z","lastTransitionTime":"2025-12-03T10:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.033399 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.033721 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.033860 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.034066 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.034280 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:10Z","lastTransitionTime":"2025-12-03T10:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.138490 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.138914 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.139149 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.139378 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.139537 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:10Z","lastTransitionTime":"2025-12-03T10:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.243101 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.243185 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.243204 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.243229 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.243247 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:10Z","lastTransitionTime":"2025-12-03T10:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.345818 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.346163 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.346320 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.346529 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.346679 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:10Z","lastTransitionTime":"2025-12-03T10:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.449984 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.450464 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.450641 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.450841 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.450990 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:10Z","lastTransitionTime":"2025-12-03T10:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.554328 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.554413 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.554436 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.554460 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.554479 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:10Z","lastTransitionTime":"2025-12-03T10:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.657586 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.657679 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.657699 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.657726 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.657744 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:10Z","lastTransitionTime":"2025-12-03T10:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.760867 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.760911 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.760924 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.760946 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.760960 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:10Z","lastTransitionTime":"2025-12-03T10:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.847849 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:10 crc kubenswrapper[4646]: E1203 10:55:10.848076 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.864289 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.864406 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.864423 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.864446 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.864463 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:10Z","lastTransitionTime":"2025-12-03T10:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.968487 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.968602 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.968672 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.968712 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:10 crc kubenswrapper[4646]: I1203 10:55:10.968787 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:10Z","lastTransitionTime":"2025-12-03T10:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.072259 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.072377 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.072408 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.072440 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.072465 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:11Z","lastTransitionTime":"2025-12-03T10:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.175285 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.175382 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.175401 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.175426 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.175444 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:11Z","lastTransitionTime":"2025-12-03T10:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.277787 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.277842 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.277859 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.277882 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.277898 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:11Z","lastTransitionTime":"2025-12-03T10:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.381466 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.381531 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.381554 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.381586 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.381610 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:11Z","lastTransitionTime":"2025-12-03T10:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.485926 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.486193 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.486223 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.486250 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.486272 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:11Z","lastTransitionTime":"2025-12-03T10:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.589826 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.589890 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.589907 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.589934 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.589951 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:11Z","lastTransitionTime":"2025-12-03T10:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.692131 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.692196 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.692224 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.692248 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.692265 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:11Z","lastTransitionTime":"2025-12-03T10:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.795522 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.795565 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.795576 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.795594 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.795606 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:11Z","lastTransitionTime":"2025-12-03T10:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.848260 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.848374 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.848425 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:11 crc kubenswrapper[4646]: E1203 10:55:11.848518 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:11 crc kubenswrapper[4646]: E1203 10:55:11.848766 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:11 crc kubenswrapper[4646]: E1203 10:55:11.848913 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.898541 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.898611 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.898658 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.898689 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:11 crc kubenswrapper[4646]: I1203 10:55:11.898710 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:11Z","lastTransitionTime":"2025-12-03T10:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.003449 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.003504 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.003521 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.003545 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.003563 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:12Z","lastTransitionTime":"2025-12-03T10:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.107114 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.107192 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.107217 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.107251 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.107277 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:12Z","lastTransitionTime":"2025-12-03T10:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.210544 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.210619 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.210638 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.210665 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.210685 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:12Z","lastTransitionTime":"2025-12-03T10:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.313942 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.314018 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.314038 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.314071 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.314097 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:12Z","lastTransitionTime":"2025-12-03T10:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.418719 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.418775 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.418792 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.418816 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.418836 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:12Z","lastTransitionTime":"2025-12-03T10:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.521961 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.522022 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.522041 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.522069 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.522093 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:12Z","lastTransitionTime":"2025-12-03T10:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.624810 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.624890 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.624913 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.624945 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.624967 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:12Z","lastTransitionTime":"2025-12-03T10:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.727938 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.728018 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.728038 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.728062 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.728081 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:12Z","lastTransitionTime":"2025-12-03T10:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.830953 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.831025 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.831048 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.831083 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.831101 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:12Z","lastTransitionTime":"2025-12-03T10:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.847329 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:12 crc kubenswrapper[4646]: E1203 10:55:12.847499 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.866940 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.934213 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.934250 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.934262 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.934281 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:12 crc kubenswrapper[4646]: I1203 10:55:12.934293 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:12Z","lastTransitionTime":"2025-12-03T10:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.036983 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.037095 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.037121 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.037149 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.037170 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:13Z","lastTransitionTime":"2025-12-03T10:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.140173 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.140235 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.140253 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.140278 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.140295 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:13Z","lastTransitionTime":"2025-12-03T10:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.244278 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.244385 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.244410 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.244440 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.244462 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:13Z","lastTransitionTime":"2025-12-03T10:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.346730 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.346779 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.346792 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.346812 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.346827 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:13Z","lastTransitionTime":"2025-12-03T10:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.450595 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.450652 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.450667 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.450689 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.450704 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:13Z","lastTransitionTime":"2025-12-03T10:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.553757 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.553813 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.553831 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.553854 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.553871 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:13Z","lastTransitionTime":"2025-12-03T10:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.657604 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.657670 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.657682 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.657701 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.657716 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:13Z","lastTransitionTime":"2025-12-03T10:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.759653 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.759704 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.759715 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.759734 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.759745 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:13Z","lastTransitionTime":"2025-12-03T10:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.847777 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.847897 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.847917 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:13 crc kubenswrapper[4646]: E1203 10:55:13.848060 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:13 crc kubenswrapper[4646]: E1203 10:55:13.848154 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:13 crc kubenswrapper[4646]: E1203 10:55:13.848891 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.849286 4646 scope.go:117] "RemoveContainer" containerID="8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034" Dec 03 10:55:13 crc kubenswrapper[4646]: E1203 10:55:13.849606 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.863390 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.863453 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.863474 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.863501 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.863527 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:13Z","lastTransitionTime":"2025-12-03T10:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.894123 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.8941001910000002 podStartE2EDuration="1.894100191s" podCreationTimestamp="2025-12-03 10:55:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:55:13.892844105 +0000 UTC m=+90.355900300" watchObservedRunningTime="2025-12-03 10:55:13.894100191 +0000 UTC m=+90.357156326" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.922978 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gprb8" podStartSLOduration=71.922963381 podStartE2EDuration="1m11.922963381s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:55:13.921903351 +0000 UTC m=+90.384959536" watchObservedRunningTime="2025-12-03 10:55:13.922963381 +0000 UTC m=+90.386019506" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.966498 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.966539 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.966549 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.966569 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:13 crc kubenswrapper[4646]: I1203 10:55:13.966581 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:13Z","lastTransitionTime":"2025-12-03T10:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.039669 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-ffm54" podStartSLOduration=73.039648215 podStartE2EDuration="1m13.039648215s" podCreationTimestamp="2025-12-03 10:54:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:55:14.025005734 +0000 UTC m=+90.488061869" watchObservedRunningTime="2025-12-03 10:55:14.039648215 +0000 UTC m=+90.502704360" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.068800 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.068840 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.068848 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.068863 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.068872 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:14Z","lastTransitionTime":"2025-12-03T10:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.081014 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-lhrsz" podStartSLOduration=72.080993415 podStartE2EDuration="1m12.080993415s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:55:14.053910365 +0000 UTC m=+90.516966500" watchObservedRunningTime="2025-12-03 10:55:14.080993415 +0000 UTC m=+90.544049550" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.092165 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=71.092133478 podStartE2EDuration="1m11.092133478s" podCreationTimestamp="2025-12-03 10:54:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:55:14.080618814 +0000 UTC m=+90.543674959" watchObservedRunningTime="2025-12-03 10:55:14.092133478 +0000 UTC m=+90.555189653" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.093429 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=10.093409133 podStartE2EDuration="10.093409133s" podCreationTimestamp="2025-12-03 10:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:55:14.092388755 +0000 UTC m=+90.555444890" watchObservedRunningTime="2025-12-03 10:55:14.093409133 +0000 UTC m=+90.556465348" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.122578 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=73.122554061 podStartE2EDuration="1m13.122554061s" podCreationTimestamp="2025-12-03 10:54:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:55:14.106629014 +0000 UTC m=+90.569685179" watchObservedRunningTime="2025-12-03 10:55:14.122554061 +0000 UTC m=+90.585610226" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.136862 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=41.136833952 podStartE2EDuration="41.136833952s" podCreationTimestamp="2025-12-03 10:54:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:55:14.12357705 +0000 UTC m=+90.586633195" watchObservedRunningTime="2025-12-03 10:55:14.136833952 +0000 UTC m=+90.599890087" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.151611 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-cbp2x" podStartSLOduration=72.151594896 podStartE2EDuration="1m12.151594896s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:55:14.150565297 +0000 UTC m=+90.613621472" watchObservedRunningTime="2025-12-03 10:55:14.151594896 +0000 UTC m=+90.614651031" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.164764 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podStartSLOduration=72.164740655 podStartE2EDuration="1m12.164740655s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:55:14.164291952 +0000 UTC m=+90.627348127" watchObservedRunningTime="2025-12-03 10:55:14.164740655 +0000 UTC m=+90.627796820" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.171117 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.171165 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.171182 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.171205 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.171222 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:14Z","lastTransitionTime":"2025-12-03T10:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.209811 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-b49xq" podStartSLOduration=72.209792969 podStartE2EDuration="1m12.209792969s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:55:14.209265504 +0000 UTC m=+90.672321639" watchObservedRunningTime="2025-12-03 10:55:14.209792969 +0000 UTC m=+90.672849104" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.274143 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.274187 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.274196 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.274212 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.274221 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:14Z","lastTransitionTime":"2025-12-03T10:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.377989 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.378075 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.378102 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.378135 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.378159 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:14Z","lastTransitionTime":"2025-12-03T10:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.481533 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.481893 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.482210 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.482587 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.482894 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:14Z","lastTransitionTime":"2025-12-03T10:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.586383 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.586449 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.586470 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.586499 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.586522 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:14Z","lastTransitionTime":"2025-12-03T10:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.689238 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.689292 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.689309 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.689487 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.689513 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:14Z","lastTransitionTime":"2025-12-03T10:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.792972 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.793375 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.793528 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.793829 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.793998 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:14Z","lastTransitionTime":"2025-12-03T10:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.847376 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:14 crc kubenswrapper[4646]: E1203 10:55:14.848027 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.897602 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.897862 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.897993 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.898162 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:14 crc kubenswrapper[4646]: I1203 10:55:14.898309 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:14Z","lastTransitionTime":"2025-12-03T10:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.002586 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.003641 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.003990 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.004268 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.004567 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:15Z","lastTransitionTime":"2025-12-03T10:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.108219 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.108612 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.108865 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.109082 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.109308 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:15Z","lastTransitionTime":"2025-12-03T10:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.212088 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.212165 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.212192 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.212221 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.212238 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:15Z","lastTransitionTime":"2025-12-03T10:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.316414 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.316805 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.317051 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.317224 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.317419 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:15Z","lastTransitionTime":"2025-12-03T10:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.420655 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.421117 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.421403 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.421891 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.422312 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:15Z","lastTransitionTime":"2025-12-03T10:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.526067 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.526139 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.526161 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.526188 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.526207 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:15Z","lastTransitionTime":"2025-12-03T10:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.629763 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.629854 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.629880 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.629909 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.629932 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:15Z","lastTransitionTime":"2025-12-03T10:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.732755 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.732811 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.732828 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.732852 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.732869 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:15Z","lastTransitionTime":"2025-12-03T10:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.835995 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.836072 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.836097 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.836132 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.836158 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:15Z","lastTransitionTime":"2025-12-03T10:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.848658 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.848698 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:15 crc kubenswrapper[4646]: E1203 10:55:15.848877 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.849179 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:15 crc kubenswrapper[4646]: E1203 10:55:15.849311 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:15 crc kubenswrapper[4646]: E1203 10:55:15.849608 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.939959 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.940026 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.940188 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.940215 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:15 crc kubenswrapper[4646]: I1203 10:55:15.940232 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:15Z","lastTransitionTime":"2025-12-03T10:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.043745 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.043808 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.043832 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.043867 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.043890 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:16Z","lastTransitionTime":"2025-12-03T10:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.147263 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.147312 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.147329 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.147396 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.147417 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:16Z","lastTransitionTime":"2025-12-03T10:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.249551 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.249602 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.249619 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.249645 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.249664 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:16Z","lastTransitionTime":"2025-12-03T10:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.353314 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.353429 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.353464 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.353497 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.353519 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:16Z","lastTransitionTime":"2025-12-03T10:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.456774 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.456834 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.456851 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.456876 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.456895 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:16Z","lastTransitionTime":"2025-12-03T10:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.560546 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.560599 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.560616 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.560642 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.560660 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:16Z","lastTransitionTime":"2025-12-03T10:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.664469 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.664935 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.665192 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.665452 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.665633 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:16Z","lastTransitionTime":"2025-12-03T10:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.717389 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.717798 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.718137 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.718475 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.718890 4646 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T10:55:16Z","lastTransitionTime":"2025-12-03T10:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.787687 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq"] Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.788102 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.790452 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.791398 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.791399 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.795381 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.848255 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:16 crc kubenswrapper[4646]: E1203 10:55:16.848417 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.873267 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g2dpq\" (UID: \"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.873362 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g2dpq\" (UID: \"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.873443 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g2dpq\" (UID: \"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.873527 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g2dpq\" (UID: \"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.873827 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g2dpq\" (UID: \"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.974292 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g2dpq\" (UID: \"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.974371 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g2dpq\" (UID: \"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.974412 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g2dpq\" (UID: \"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.974429 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g2dpq\" (UID: \"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.974486 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g2dpq\" (UID: \"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.974524 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g2dpq\" (UID: \"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.974554 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g2dpq\" (UID: \"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.976024 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g2dpq\" (UID: \"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.980203 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g2dpq\" (UID: \"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:16 crc kubenswrapper[4646]: I1203 10:55:16.991892 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g2dpq\" (UID: \"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:17 crc kubenswrapper[4646]: I1203 10:55:17.102979 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" Dec 03 10:55:17 crc kubenswrapper[4646]: I1203 10:55:17.440913 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" event={"ID":"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b","Type":"ContainerStarted","Data":"5793b2ecfc1dd5adf73fff4cf817ebb0e8d0db62aed8f0963f34344c1e9472f7"} Dec 03 10:55:17 crc kubenswrapper[4646]: I1203 10:55:17.441221 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" event={"ID":"9f0f32a8-b46f-4b5e-a289-cd0d1e5cb17b","Type":"ContainerStarted","Data":"90f1d92c33c33533683529e559e84af492d9ea334b8d922b2fdb0e734cea79a0"} Dec 03 10:55:17 crc kubenswrapper[4646]: I1203 10:55:17.466006 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g2dpq" podStartSLOduration=75.465980797 podStartE2EDuration="1m15.465980797s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:55:17.463673233 +0000 UTC m=+93.926729378" watchObservedRunningTime="2025-12-03 10:55:17.465980797 +0000 UTC m=+93.929036962" Dec 03 10:55:17 crc kubenswrapper[4646]: I1203 10:55:17.847291 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:17 crc kubenswrapper[4646]: I1203 10:55:17.847424 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:17 crc kubenswrapper[4646]: E1203 10:55:17.847605 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:17 crc kubenswrapper[4646]: I1203 10:55:17.847632 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:17 crc kubenswrapper[4646]: E1203 10:55:17.847837 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:17 crc kubenswrapper[4646]: E1203 10:55:17.847998 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:18 crc kubenswrapper[4646]: I1203 10:55:18.847663 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:18 crc kubenswrapper[4646]: E1203 10:55:18.848216 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:19 crc kubenswrapper[4646]: I1203 10:55:19.847680 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:19 crc kubenswrapper[4646]: E1203 10:55:19.847830 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:19 crc kubenswrapper[4646]: I1203 10:55:19.847680 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:19 crc kubenswrapper[4646]: E1203 10:55:19.848031 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:19 crc kubenswrapper[4646]: I1203 10:55:19.847873 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:19 crc kubenswrapper[4646]: E1203 10:55:19.848149 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:20 crc kubenswrapper[4646]: I1203 10:55:20.413669 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs\") pod \"network-metrics-daemon-rs5rq\" (UID: \"51b171e4-cc29-4118-908c-d13825362e1c\") " pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:20 crc kubenswrapper[4646]: E1203 10:55:20.413893 4646 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:55:20 crc kubenswrapper[4646]: E1203 10:55:20.413994 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs podName:51b171e4-cc29-4118-908c-d13825362e1c nodeName:}" failed. No retries permitted until 2025-12-03 10:56:24.413969479 +0000 UTC m=+160.877025654 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs") pod "network-metrics-daemon-rs5rq" (UID: "51b171e4-cc29-4118-908c-d13825362e1c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 10:55:20 crc kubenswrapper[4646]: I1203 10:55:20.847889 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:20 crc kubenswrapper[4646]: E1203 10:55:20.848162 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:22 crc kubenswrapper[4646]: I1203 10:55:22.026911 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:22 crc kubenswrapper[4646]: I1203 10:55:22.027085 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:22 crc kubenswrapper[4646]: I1203 10:55:22.027198 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:22 crc kubenswrapper[4646]: E1203 10:55:22.027309 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:22 crc kubenswrapper[4646]: E1203 10:55:22.027197 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:22 crc kubenswrapper[4646]: E1203 10:55:22.027488 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:22 crc kubenswrapper[4646]: I1203 10:55:22.848298 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:22 crc kubenswrapper[4646]: E1203 10:55:22.848506 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:23 crc kubenswrapper[4646]: I1203 10:55:23.848103 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:23 crc kubenswrapper[4646]: E1203 10:55:23.848324 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:23 crc kubenswrapper[4646]: I1203 10:55:23.848390 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:23 crc kubenswrapper[4646]: I1203 10:55:23.848744 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:23 crc kubenswrapper[4646]: E1203 10:55:23.850473 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:23 crc kubenswrapper[4646]: E1203 10:55:23.850569 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:24 crc kubenswrapper[4646]: I1203 10:55:24.847516 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:24 crc kubenswrapper[4646]: E1203 10:55:24.847626 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:25 crc kubenswrapper[4646]: I1203 10:55:25.847691 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:25 crc kubenswrapper[4646]: I1203 10:55:25.847738 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:25 crc kubenswrapper[4646]: I1203 10:55:25.847771 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:25 crc kubenswrapper[4646]: E1203 10:55:25.847894 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:25 crc kubenswrapper[4646]: E1203 10:55:25.848078 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:25 crc kubenswrapper[4646]: E1203 10:55:25.848256 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:26 crc kubenswrapper[4646]: I1203 10:55:26.848217 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:26 crc kubenswrapper[4646]: E1203 10:55:26.848477 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:26 crc kubenswrapper[4646]: I1203 10:55:26.849795 4646 scope.go:117] "RemoveContainer" containerID="8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034" Dec 03 10:55:26 crc kubenswrapper[4646]: E1203 10:55:26.850126 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9lfpq_openshift-ovn-kubernetes(8cd1981e-44c4-4251-81d1-c07c9d45258b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" Dec 03 10:55:27 crc kubenswrapper[4646]: I1203 10:55:27.848198 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:27 crc kubenswrapper[4646]: I1203 10:55:27.848237 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:27 crc kubenswrapper[4646]: I1203 10:55:27.848225 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:27 crc kubenswrapper[4646]: E1203 10:55:27.848455 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:27 crc kubenswrapper[4646]: E1203 10:55:27.848562 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:27 crc kubenswrapper[4646]: E1203 10:55:27.848699 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:28 crc kubenswrapper[4646]: I1203 10:55:28.847637 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:28 crc kubenswrapper[4646]: E1203 10:55:28.847874 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:29 crc kubenswrapper[4646]: I1203 10:55:29.848099 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:29 crc kubenswrapper[4646]: I1203 10:55:29.848197 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:29 crc kubenswrapper[4646]: E1203 10:55:29.848385 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:29 crc kubenswrapper[4646]: E1203 10:55:29.849591 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:29 crc kubenswrapper[4646]: I1203 10:55:29.848134 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:29 crc kubenswrapper[4646]: E1203 10:55:29.849734 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:30 crc kubenswrapper[4646]: I1203 10:55:30.847474 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:30 crc kubenswrapper[4646]: E1203 10:55:30.848091 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:31 crc kubenswrapper[4646]: I1203 10:55:31.847916 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:31 crc kubenswrapper[4646]: E1203 10:55:31.848724 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:31 crc kubenswrapper[4646]: I1203 10:55:31.848279 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:31 crc kubenswrapper[4646]: E1203 10:55:31.849008 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:31 crc kubenswrapper[4646]: I1203 10:55:31.848039 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:31 crc kubenswrapper[4646]: E1203 10:55:31.849309 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:32 crc kubenswrapper[4646]: I1203 10:55:32.847795 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:32 crc kubenswrapper[4646]: E1203 10:55:32.847909 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:33 crc kubenswrapper[4646]: I1203 10:55:33.847971 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:33 crc kubenswrapper[4646]: I1203 10:55:33.848050 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:33 crc kubenswrapper[4646]: I1203 10:55:33.848124 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:33 crc kubenswrapper[4646]: E1203 10:55:33.850025 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:33 crc kubenswrapper[4646]: E1203 10:55:33.850234 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:33 crc kubenswrapper[4646]: E1203 10:55:33.850497 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:34 crc kubenswrapper[4646]: I1203 10:55:34.848279 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:34 crc kubenswrapper[4646]: E1203 10:55:34.848930 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:35 crc kubenswrapper[4646]: I1203 10:55:35.519423 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cbp2x_0b003d65-42ef-4a54-96e8-33428e114a1d/kube-multus/1.log" Dec 03 10:55:35 crc kubenswrapper[4646]: I1203 10:55:35.520332 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cbp2x_0b003d65-42ef-4a54-96e8-33428e114a1d/kube-multus/0.log" Dec 03 10:55:35 crc kubenswrapper[4646]: I1203 10:55:35.520475 4646 generic.go:334] "Generic (PLEG): container finished" podID="0b003d65-42ef-4a54-96e8-33428e114a1d" containerID="ffa715e7e498fda77dee5e1e92dece871ede3ada1b73f41f8f35ff15fc8910fd" exitCode=1 Dec 03 10:55:35 crc kubenswrapper[4646]: I1203 10:55:35.520532 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cbp2x" event={"ID":"0b003d65-42ef-4a54-96e8-33428e114a1d","Type":"ContainerDied","Data":"ffa715e7e498fda77dee5e1e92dece871ede3ada1b73f41f8f35ff15fc8910fd"} Dec 03 10:55:35 crc kubenswrapper[4646]: I1203 10:55:35.520597 4646 scope.go:117] "RemoveContainer" containerID="21517adcd5c8e26fc193d3b64e205fc38edac5e5a3918c81bce5d252b52abe4f" Dec 03 10:55:35 crc kubenswrapper[4646]: I1203 10:55:35.521160 4646 scope.go:117] "RemoveContainer" containerID="ffa715e7e498fda77dee5e1e92dece871ede3ada1b73f41f8f35ff15fc8910fd" Dec 03 10:55:35 crc kubenswrapper[4646]: E1203 10:55:35.521490 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-cbp2x_openshift-multus(0b003d65-42ef-4a54-96e8-33428e114a1d)\"" pod="openshift-multus/multus-cbp2x" podUID="0b003d65-42ef-4a54-96e8-33428e114a1d" Dec 03 10:55:35 crc kubenswrapper[4646]: I1203 10:55:35.847726 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:35 crc kubenswrapper[4646]: I1203 10:55:35.847794 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:35 crc kubenswrapper[4646]: E1203 10:55:35.847927 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:35 crc kubenswrapper[4646]: I1203 10:55:35.848170 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:35 crc kubenswrapper[4646]: E1203 10:55:35.848270 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:35 crc kubenswrapper[4646]: E1203 10:55:35.848529 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:36 crc kubenswrapper[4646]: I1203 10:55:36.525909 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cbp2x_0b003d65-42ef-4a54-96e8-33428e114a1d/kube-multus/1.log" Dec 03 10:55:36 crc kubenswrapper[4646]: I1203 10:55:36.847996 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:36 crc kubenswrapper[4646]: E1203 10:55:36.848168 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:37 crc kubenswrapper[4646]: I1203 10:55:37.847655 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:37 crc kubenswrapper[4646]: I1203 10:55:37.847765 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:37 crc kubenswrapper[4646]: I1203 10:55:37.847830 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:37 crc kubenswrapper[4646]: E1203 10:55:37.847863 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:37 crc kubenswrapper[4646]: E1203 10:55:37.848047 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:37 crc kubenswrapper[4646]: E1203 10:55:37.848262 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:38 crc kubenswrapper[4646]: I1203 10:55:38.848245 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:38 crc kubenswrapper[4646]: E1203 10:55:38.848525 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:39 crc kubenswrapper[4646]: I1203 10:55:39.847748 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:39 crc kubenswrapper[4646]: I1203 10:55:39.847794 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:39 crc kubenswrapper[4646]: I1203 10:55:39.847848 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:39 crc kubenswrapper[4646]: E1203 10:55:39.847975 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:39 crc kubenswrapper[4646]: E1203 10:55:39.848215 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:39 crc kubenswrapper[4646]: E1203 10:55:39.848431 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:40 crc kubenswrapper[4646]: I1203 10:55:40.847586 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:40 crc kubenswrapper[4646]: E1203 10:55:40.847745 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:41 crc kubenswrapper[4646]: I1203 10:55:41.848056 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:41 crc kubenswrapper[4646]: E1203 10:55:41.848212 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:41 crc kubenswrapper[4646]: I1203 10:55:41.848312 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:41 crc kubenswrapper[4646]: I1203 10:55:41.848479 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:41 crc kubenswrapper[4646]: E1203 10:55:41.848768 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:41 crc kubenswrapper[4646]: E1203 10:55:41.848879 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:41 crc kubenswrapper[4646]: I1203 10:55:41.849130 4646 scope.go:117] "RemoveContainer" containerID="8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034" Dec 03 10:55:42 crc kubenswrapper[4646]: I1203 10:55:42.549434 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovnkube-controller/3.log" Dec 03 10:55:42 crc kubenswrapper[4646]: I1203 10:55:42.552463 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerStarted","Data":"a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214"} Dec 03 10:55:42 crc kubenswrapper[4646]: I1203 10:55:42.553036 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:55:42 crc kubenswrapper[4646]: I1203 10:55:42.663985 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podStartSLOduration=100.663962087 podStartE2EDuration="1m40.663962087s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:55:42.58174952 +0000 UTC m=+119.044805675" watchObservedRunningTime="2025-12-03 10:55:42.663962087 +0000 UTC m=+119.127018242" Dec 03 10:55:42 crc kubenswrapper[4646]: I1203 10:55:42.665204 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-rs5rq"] Dec 03 10:55:42 crc kubenswrapper[4646]: I1203 10:55:42.665346 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:42 crc kubenswrapper[4646]: E1203 10:55:42.665449 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:42 crc kubenswrapper[4646]: I1203 10:55:42.847458 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:42 crc kubenswrapper[4646]: E1203 10:55:42.847776 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:43 crc kubenswrapper[4646]: E1203 10:55:43.777872 4646 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 03 10:55:43 crc kubenswrapper[4646]: I1203 10:55:43.847922 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:43 crc kubenswrapper[4646]: E1203 10:55:43.849829 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:43 crc kubenswrapper[4646]: I1203 10:55:43.849917 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:43 crc kubenswrapper[4646]: E1203 10:55:43.850084 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:43 crc kubenswrapper[4646]: E1203 10:55:43.933470 4646 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 10:55:44 crc kubenswrapper[4646]: I1203 10:55:44.847806 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:44 crc kubenswrapper[4646]: E1203 10:55:44.848016 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:44 crc kubenswrapper[4646]: I1203 10:55:44.849286 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:44 crc kubenswrapper[4646]: E1203 10:55:44.849676 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:45 crc kubenswrapper[4646]: I1203 10:55:45.848252 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:45 crc kubenswrapper[4646]: I1203 10:55:45.848258 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:45 crc kubenswrapper[4646]: E1203 10:55:45.848484 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:45 crc kubenswrapper[4646]: E1203 10:55:45.848649 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:46 crc kubenswrapper[4646]: I1203 10:55:46.847801 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:46 crc kubenswrapper[4646]: I1203 10:55:46.847857 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:46 crc kubenswrapper[4646]: E1203 10:55:46.848014 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:46 crc kubenswrapper[4646]: E1203 10:55:46.848223 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:47 crc kubenswrapper[4646]: I1203 10:55:47.847963 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:47 crc kubenswrapper[4646]: I1203 10:55:47.848000 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:47 crc kubenswrapper[4646]: E1203 10:55:47.848192 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:47 crc kubenswrapper[4646]: E1203 10:55:47.848379 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:47 crc kubenswrapper[4646]: I1203 10:55:47.848904 4646 scope.go:117] "RemoveContainer" containerID="ffa715e7e498fda77dee5e1e92dece871ede3ada1b73f41f8f35ff15fc8910fd" Dec 03 10:55:48 crc kubenswrapper[4646]: I1203 10:55:48.577757 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cbp2x_0b003d65-42ef-4a54-96e8-33428e114a1d/kube-multus/1.log" Dec 03 10:55:48 crc kubenswrapper[4646]: I1203 10:55:48.578124 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cbp2x" event={"ID":"0b003d65-42ef-4a54-96e8-33428e114a1d","Type":"ContainerStarted","Data":"c364a629d49eff8128fc43ea166baeec2f873dc2d1712a4bbfc1c90a0cb2ec4c"} Dec 03 10:55:48 crc kubenswrapper[4646]: I1203 10:55:48.848280 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:48 crc kubenswrapper[4646]: E1203 10:55:48.848500 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:48 crc kubenswrapper[4646]: I1203 10:55:48.848288 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:48 crc kubenswrapper[4646]: E1203 10:55:48.848772 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:48 crc kubenswrapper[4646]: E1203 10:55:48.934185 4646 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 10:55:49 crc kubenswrapper[4646]: I1203 10:55:49.848304 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:49 crc kubenswrapper[4646]: E1203 10:55:49.848493 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:49 crc kubenswrapper[4646]: I1203 10:55:49.848761 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:49 crc kubenswrapper[4646]: E1203 10:55:49.848851 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:50 crc kubenswrapper[4646]: I1203 10:55:50.848013 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:50 crc kubenswrapper[4646]: I1203 10:55:50.848061 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:50 crc kubenswrapper[4646]: E1203 10:55:50.848212 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:50 crc kubenswrapper[4646]: E1203 10:55:50.848364 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:51 crc kubenswrapper[4646]: I1203 10:55:51.847730 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:51 crc kubenswrapper[4646]: I1203 10:55:51.847808 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:51 crc kubenswrapper[4646]: E1203 10:55:51.847917 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:51 crc kubenswrapper[4646]: E1203 10:55:51.848053 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:52 crc kubenswrapper[4646]: I1203 10:55:52.847409 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:52 crc kubenswrapper[4646]: I1203 10:55:52.847423 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:52 crc kubenswrapper[4646]: E1203 10:55:52.847676 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs5rq" podUID="51b171e4-cc29-4118-908c-d13825362e1c" Dec 03 10:55:52 crc kubenswrapper[4646]: E1203 10:55:52.847882 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 10:55:53 crc kubenswrapper[4646]: I1203 10:55:53.851591 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:53 crc kubenswrapper[4646]: E1203 10:55:53.852675 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 10:55:53 crc kubenswrapper[4646]: I1203 10:55:53.852703 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:53 crc kubenswrapper[4646]: E1203 10:55:53.852800 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 10:55:54 crc kubenswrapper[4646]: I1203 10:55:54.847705 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:55:54 crc kubenswrapper[4646]: I1203 10:55:54.847785 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:55:54 crc kubenswrapper[4646]: I1203 10:55:54.852026 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 10:55:54 crc kubenswrapper[4646]: I1203 10:55:54.852057 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 10:55:54 crc kubenswrapper[4646]: I1203 10:55:54.852449 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 10:55:54 crc kubenswrapper[4646]: I1203 10:55:54.852742 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 10:55:55 crc kubenswrapper[4646]: I1203 10:55:55.848288 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:55:55 crc kubenswrapper[4646]: I1203 10:55:55.848290 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:55:55 crc kubenswrapper[4646]: I1203 10:55:55.850944 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 10:55:55 crc kubenswrapper[4646]: I1203 10:55:55.851678 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.255468 4646 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.302358 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.303011 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.308521 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.315436 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.318800 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.319046 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.319223 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.319441 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.319719 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.319764 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.319869 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.327101 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.327206 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.327409 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.327759 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.327822 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.327999 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.328096 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.328538 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.329868 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-czl94"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.330710 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.330895 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q7bzw"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.331941 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.332480 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.333032 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.333893 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.348149 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.349302 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.349574 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.349611 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.349718 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.349775 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.349898 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.349940 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.350023 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.350123 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.350164 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.350126 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.350243 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.350285 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.350166 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.349725 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.350440 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.350503 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.350549 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.350684 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.350760 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-lcfx6"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.351367 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.352322 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.352790 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.353221 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2trtv"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.353859 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.355104 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.357173 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nb6wv"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.357669 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.358596 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.359294 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.360715 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.363422 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.363722 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.363835 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.367573 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-697vl"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.367936 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.368063 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.368328 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sc5lh"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.368797 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sc5lh" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.369021 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.369309 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.375486 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.376036 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.376077 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.376139 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.376277 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.376296 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.376387 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.381778 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.381935 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.382617 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.383518 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.383880 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.384533 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.385123 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.386041 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.386197 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-lcfx6"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.386263 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.386501 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.386749 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.386766 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.386905 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.387056 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.387162 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2trtv"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.387209 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.388106 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.388352 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.388504 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.388620 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.388853 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.392498 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.392768 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.393986 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.394700 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.396122 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.401744 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.401801 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.401747 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.402017 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.401806 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.403116 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.405100 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q7bzw"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.424174 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.426324 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.426383 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.426402 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-rvbtr"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.426939 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.434883 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-czl94"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.434929 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-52ndk"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.435360 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.442228 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.443606 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.443873 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-697vl"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.444239 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-l2kck"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.444582 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.444710 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-l2kck" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445131 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445276 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445419 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445556 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445585 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8936a9d-f33c-4de3-a759-7df7cf3e48b2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-g2szh\" (UID: \"c8936a9d-f33c-4de3-a759-7df7cf3e48b2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445608 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a0dc011-d040-4ec5-babf-d95b03ec699e-config\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445629 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a0dc011-d040-4ec5-babf-d95b03ec699e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445650 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-encryption-config\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445668 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blsgz\" (UniqueName: \"kubernetes.io/projected/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-kube-api-access-blsgz\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445684 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1cc84826-c569-408e-94e9-8cf9bc49eae9-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-lcfx6\" (UID: \"1cc84826-c569-408e-94e9-8cf9bc49eae9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445703 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445719 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wnzv\" (UniqueName: \"kubernetes.io/projected/e51eada8-802d-4064-8705-2647ffa44189-kube-api-access-8wnzv\") pod \"controller-manager-879f6c89f-czl94\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445722 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445740 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e51eada8-802d-4064-8705-2647ffa44189-serving-cert\") pod \"controller-manager-879f6c89f-czl94\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445766 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-etcd-client\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445783 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d722\" (UniqueName: \"kubernetes.io/projected/aa151fb7-08d1-48b4-b477-e3310ff2274d-kube-api-access-8d722\") pod \"authentication-operator-69f744f599-2trtv\" (UID: \"aa151fb7-08d1-48b4-b477-e3310ff2274d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445800 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445814 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9a0dc011-d040-4ec5-babf-d95b03ec699e-etcd-serving-ca\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445832 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-czl94\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445845 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9a0dc011-d040-4ec5-babf-d95b03ec699e-etcd-client\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445863 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80cf56c0-7055-4e98-9fad-f728dd8b992a-client-ca\") pod \"route-controller-manager-6576b87f9c-plqjj\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445880 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cc84826-c569-408e-94e9-8cf9bc49eae9-config\") pod \"machine-api-operator-5694c8668f-lcfx6\" (UID: \"1cc84826-c569-408e-94e9-8cf9bc49eae9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445895 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9a0dc011-d040-4ec5-babf-d95b03ec699e-node-pullsecrets\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445910 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8936a9d-f33c-4de3-a759-7df7cf3e48b2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-g2szh\" (UID: \"c8936a9d-f33c-4de3-a759-7df7cf3e48b2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445926 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68dk6\" (UniqueName: \"kubernetes.io/projected/c8936a9d-f33c-4de3-a759-7df7cf3e48b2-kube-api-access-68dk6\") pod \"openshift-apiserver-operator-796bbdcf4f-g2szh\" (UID: \"c8936a9d-f33c-4de3-a759-7df7cf3e48b2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445942 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-client-ca\") pod \"controller-manager-879f6c89f-czl94\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445960 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9a0dc011-d040-4ec5-babf-d95b03ec699e-image-import-ca\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445977 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f653a1d6-4938-418c-9b40-ad23d437a156-machine-approver-tls\") pod \"machine-approver-56656f9798-p7xvl\" (UID: \"f653a1d6-4938-418c-9b40-ad23d437a156\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445992 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn2fx\" (UniqueName: \"kubernetes.io/projected/f653a1d6-4938-418c-9b40-ad23d437a156-kube-api-access-hn2fx\") pod \"machine-approver-56656f9798-p7xvl\" (UID: \"f653a1d6-4938-418c-9b40-ad23d437a156\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446008 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9a0dc011-d040-4ec5-babf-d95b03ec699e-encryption-config\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446026 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9a0dc011-d040-4ec5-babf-d95b03ec699e-audit-dir\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446070 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-serving-cert\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446095 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-audit-dir\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446110 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f653a1d6-4938-418c-9b40-ad23d437a156-config\") pod \"machine-approver-56656f9798-p7xvl\" (UID: \"f653a1d6-4938-418c-9b40-ad23d437a156\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446144 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa151fb7-08d1-48b4-b477-e3310ff2274d-config\") pod \"authentication-operator-69f744f599-2trtv\" (UID: \"aa151fb7-08d1-48b4-b477-e3310ff2274d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.445880 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sc5lh"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446185 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-config\") pod \"controller-manager-879f6c89f-czl94\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446366 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1cc84826-c569-408e-94e9-8cf9bc49eae9-images\") pod \"machine-api-operator-5694c8668f-lcfx6\" (UID: \"1cc84826-c569-408e-94e9-8cf9bc49eae9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446397 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80cf56c0-7055-4e98-9fad-f728dd8b992a-serving-cert\") pod \"route-controller-manager-6576b87f9c-plqjj\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446403 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446420 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9a0dc011-d040-4ec5-babf-d95b03ec699e-audit\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446595 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a0dc011-d040-4ec5-babf-d95b03ec699e-serving-cert\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446626 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa151fb7-08d1-48b4-b477-e3310ff2274d-serving-cert\") pod \"authentication-operator-69f744f599-2trtv\" (UID: \"aa151fb7-08d1-48b4-b477-e3310ff2274d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446656 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-audit-policies\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446677 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa151fb7-08d1-48b4-b477-e3310ff2274d-service-ca-bundle\") pod \"authentication-operator-69f744f599-2trtv\" (UID: \"aa151fb7-08d1-48b4-b477-e3310ff2274d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446698 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f653a1d6-4938-418c-9b40-ad23d437a156-auth-proxy-config\") pod \"machine-approver-56656f9798-p7xvl\" (UID: \"f653a1d6-4938-418c-9b40-ad23d437a156\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446721 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfjx4\" (UniqueName: \"kubernetes.io/projected/1cc84826-c569-408e-94e9-8cf9bc49eae9-kube-api-access-cfjx4\") pod \"machine-api-operator-5694c8668f-lcfx6\" (UID: \"1cc84826-c569-408e-94e9-8cf9bc49eae9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446747 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-578ds\" (UniqueName: \"kubernetes.io/projected/80cf56c0-7055-4e98-9fad-f728dd8b992a-kube-api-access-578ds\") pod \"route-controller-manager-6576b87f9c-plqjj\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446767 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dq2p\" (UniqueName: \"kubernetes.io/projected/9a0dc011-d040-4ec5-babf-d95b03ec699e-kube-api-access-2dq2p\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446789 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80cf56c0-7055-4e98-9fad-f728dd8b992a-config\") pod \"route-controller-manager-6576b87f9c-plqjj\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446809 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa151fb7-08d1-48b4-b477-e3310ff2274d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2trtv\" (UID: \"aa151fb7-08d1-48b4-b477-e3310ff2274d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446819 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.446997 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.447127 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.447303 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.448068 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.448501 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nb6wv"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.448582 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.450731 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.451568 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.451802 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.451931 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.454898 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k4mmg"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.455445 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.455456 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.456429 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.456795 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.457175 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-4jgxv"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.457921 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.457955 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-4jgxv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.460990 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.478992 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.479261 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.480525 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-7whtv"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.481382 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.483758 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.484028 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.487887 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.489060 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.489766 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.491312 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.491531 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.491990 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-rvbtr"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.494409 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-4ftgs"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.495048 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.497940 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.498574 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.498582 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.500358 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.501278 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.501663 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.501872 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.502275 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.503243 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.504102 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hqnkq"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.504667 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-hqnkq" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.504955 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.506148 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.506573 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.507555 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-prp8k"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.508098 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-prp8k" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.508662 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.509834 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6rqmm"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.510196 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6rqmm" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.510759 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.511224 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.514886 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-vtndh"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.515579 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-vtndh" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.516868 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.517466 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.517540 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.519689 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mbfc8"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.520049 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-55nsv"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.520116 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.520117 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.524611 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-55nsv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.524931 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.526469 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.530232 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-mvqf6"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.530877 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-mvqf6" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.531241 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-52ndk"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.532911 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.535956 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-l2kck"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.535983 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-4jgxv"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.537852 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k4mmg"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547554 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-encryption-config\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547594 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blsgz\" (UniqueName: \"kubernetes.io/projected/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-kube-api-access-blsgz\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547610 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1cc84826-c569-408e-94e9-8cf9bc49eae9-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-lcfx6\" (UID: \"1cc84826-c569-408e-94e9-8cf9bc49eae9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547634 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547648 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wnzv\" (UniqueName: \"kubernetes.io/projected/e51eada8-802d-4064-8705-2647ffa44189-kube-api-access-8wnzv\") pod \"controller-manager-879f6c89f-czl94\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547664 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-etcd-client\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547679 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d722\" (UniqueName: \"kubernetes.io/projected/aa151fb7-08d1-48b4-b477-e3310ff2274d-kube-api-access-8d722\") pod \"authentication-operator-69f744f599-2trtv\" (UID: \"aa151fb7-08d1-48b4-b477-e3310ff2274d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547693 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e51eada8-802d-4064-8705-2647ffa44189-serving-cert\") pod \"controller-manager-879f6c89f-czl94\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547707 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547722 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9a0dc011-d040-4ec5-babf-d95b03ec699e-etcd-serving-ca\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547740 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-czl94\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547757 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9a0dc011-d040-4ec5-babf-d95b03ec699e-etcd-client\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547772 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80cf56c0-7055-4e98-9fad-f728dd8b992a-client-ca\") pod \"route-controller-manager-6576b87f9c-plqjj\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547787 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cc84826-c569-408e-94e9-8cf9bc49eae9-config\") pod \"machine-api-operator-5694c8668f-lcfx6\" (UID: \"1cc84826-c569-408e-94e9-8cf9bc49eae9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547801 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9a0dc011-d040-4ec5-babf-d95b03ec699e-node-pullsecrets\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547818 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8936a9d-f33c-4de3-a759-7df7cf3e48b2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-g2szh\" (UID: \"c8936a9d-f33c-4de3-a759-7df7cf3e48b2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547842 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68dk6\" (UniqueName: \"kubernetes.io/projected/c8936a9d-f33c-4de3-a759-7df7cf3e48b2-kube-api-access-68dk6\") pod \"openshift-apiserver-operator-796bbdcf4f-g2szh\" (UID: \"c8936a9d-f33c-4de3-a759-7df7cf3e48b2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547860 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-client-ca\") pod \"controller-manager-879f6c89f-czl94\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547881 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k68fq\" (UniqueName: \"kubernetes.io/projected/53824d3c-1dd1-4f25-b699-de9230490c84-kube-api-access-k68fq\") pod \"openshift-config-operator-7777fb866f-697vl\" (UID: \"53824d3c-1dd1-4f25-b699-de9230490c84\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547902 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9a0dc011-d040-4ec5-babf-d95b03ec699e-image-import-ca\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547923 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f653a1d6-4938-418c-9b40-ad23d437a156-machine-approver-tls\") pod \"machine-approver-56656f9798-p7xvl\" (UID: \"f653a1d6-4938-418c-9b40-ad23d437a156\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547946 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn2fx\" (UniqueName: \"kubernetes.io/projected/f653a1d6-4938-418c-9b40-ad23d437a156-kube-api-access-hn2fx\") pod \"machine-approver-56656f9798-p7xvl\" (UID: \"f653a1d6-4938-418c-9b40-ad23d437a156\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547967 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9a0dc011-d040-4ec5-babf-d95b03ec699e-encryption-config\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.547988 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9a0dc011-d040-4ec5-babf-d95b03ec699e-audit-dir\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548038 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53824d3c-1dd1-4f25-b699-de9230490c84-serving-cert\") pod \"openshift-config-operator-7777fb866f-697vl\" (UID: \"53824d3c-1dd1-4f25-b699-de9230490c84\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548057 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-serving-cert\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548072 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/53824d3c-1dd1-4f25-b699-de9230490c84-available-featuregates\") pod \"openshift-config-operator-7777fb866f-697vl\" (UID: \"53824d3c-1dd1-4f25-b699-de9230490c84\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548088 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-audit-dir\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548116 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f653a1d6-4938-418c-9b40-ad23d437a156-config\") pod \"machine-approver-56656f9798-p7xvl\" (UID: \"f653a1d6-4938-418c-9b40-ad23d437a156\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548140 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-config\") pod \"controller-manager-879f6c89f-czl94\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548154 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1cc84826-c569-408e-94e9-8cf9bc49eae9-images\") pod \"machine-api-operator-5694c8668f-lcfx6\" (UID: \"1cc84826-c569-408e-94e9-8cf9bc49eae9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548170 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa151fb7-08d1-48b4-b477-e3310ff2274d-config\") pod \"authentication-operator-69f744f599-2trtv\" (UID: \"aa151fb7-08d1-48b4-b477-e3310ff2274d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548185 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80cf56c0-7055-4e98-9fad-f728dd8b992a-serving-cert\") pod \"route-controller-manager-6576b87f9c-plqjj\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548198 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9a0dc011-d040-4ec5-babf-d95b03ec699e-audit\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548211 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a0dc011-d040-4ec5-babf-d95b03ec699e-serving-cert\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548230 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa151fb7-08d1-48b4-b477-e3310ff2274d-serving-cert\") pod \"authentication-operator-69f744f599-2trtv\" (UID: \"aa151fb7-08d1-48b4-b477-e3310ff2274d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548247 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-audit-policies\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548262 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa151fb7-08d1-48b4-b477-e3310ff2274d-service-ca-bundle\") pod \"authentication-operator-69f744f599-2trtv\" (UID: \"aa151fb7-08d1-48b4-b477-e3310ff2274d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548299 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f653a1d6-4938-418c-9b40-ad23d437a156-auth-proxy-config\") pod \"machine-approver-56656f9798-p7xvl\" (UID: \"f653a1d6-4938-418c-9b40-ad23d437a156\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548317 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfjx4\" (UniqueName: \"kubernetes.io/projected/1cc84826-c569-408e-94e9-8cf9bc49eae9-kube-api-access-cfjx4\") pod \"machine-api-operator-5694c8668f-lcfx6\" (UID: \"1cc84826-c569-408e-94e9-8cf9bc49eae9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548349 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-578ds\" (UniqueName: \"kubernetes.io/projected/80cf56c0-7055-4e98-9fad-f728dd8b992a-kube-api-access-578ds\") pod \"route-controller-manager-6576b87f9c-plqjj\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548363 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dq2p\" (UniqueName: \"kubernetes.io/projected/9a0dc011-d040-4ec5-babf-d95b03ec699e-kube-api-access-2dq2p\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548378 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80cf56c0-7055-4e98-9fad-f728dd8b992a-config\") pod \"route-controller-manager-6576b87f9c-plqjj\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548393 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa151fb7-08d1-48b4-b477-e3310ff2274d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2trtv\" (UID: \"aa151fb7-08d1-48b4-b477-e3310ff2274d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548409 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8936a9d-f33c-4de3-a759-7df7cf3e48b2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-g2szh\" (UID: \"c8936a9d-f33c-4de3-a759-7df7cf3e48b2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548422 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a0dc011-d040-4ec5-babf-d95b03ec699e-config\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.548436 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a0dc011-d040-4ec5-babf-d95b03ec699e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.549275 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a0dc011-d040-4ec5-babf-d95b03ec699e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.550957 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-czl94\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.551145 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9a0dc011-d040-4ec5-babf-d95b03ec699e-audit-dir\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.552057 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.552384 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9a0dc011-d040-4ec5-babf-d95b03ec699e-etcd-serving-ca\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.553046 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80cf56c0-7055-4e98-9fad-f728dd8b992a-client-ca\") pod \"route-controller-manager-6576b87f9c-plqjj\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.553609 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cc84826-c569-408e-94e9-8cf9bc49eae9-config\") pod \"machine-api-operator-5694c8668f-lcfx6\" (UID: \"1cc84826-c569-408e-94e9-8cf9bc49eae9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.553653 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9a0dc011-d040-4ec5-babf-d95b03ec699e-node-pullsecrets\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.554084 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8936a9d-f33c-4de3-a759-7df7cf3e48b2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-g2szh\" (UID: \"c8936a9d-f33c-4de3-a759-7df7cf3e48b2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.555907 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-audit-dir\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.576580 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80cf56c0-7055-4e98-9fad-f728dd8b992a-config\") pod \"route-controller-manager-6576b87f9c-plqjj\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.561530 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9a0dc011-d040-4ec5-babf-d95b03ec699e-image-import-ca\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.561801 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.562420 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-audit-policies\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.562426 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa151fb7-08d1-48b4-b477-e3310ff2274d-service-ca-bundle\") pod \"authentication-operator-69f744f599-2trtv\" (UID: \"aa151fb7-08d1-48b4-b477-e3310ff2274d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.562991 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f653a1d6-4938-418c-9b40-ad23d437a156-auth-proxy-config\") pod \"machine-approver-56656f9798-p7xvl\" (UID: \"f653a1d6-4938-418c-9b40-ad23d437a156\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.563109 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f653a1d6-4938-418c-9b40-ad23d437a156-config\") pod \"machine-approver-56656f9798-p7xvl\" (UID: \"f653a1d6-4938-418c-9b40-ad23d437a156\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.564483 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1cc84826-c569-408e-94e9-8cf9bc49eae9-images\") pod \"machine-api-operator-5694c8668f-lcfx6\" (UID: \"1cc84826-c569-408e-94e9-8cf9bc49eae9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.564911 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa151fb7-08d1-48b4-b477-e3310ff2274d-config\") pod \"authentication-operator-69f744f599-2trtv\" (UID: \"aa151fb7-08d1-48b4-b477-e3310ff2274d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.567893 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9a0dc011-d040-4ec5-babf-d95b03ec699e-audit\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.575224 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-config\") pod \"controller-manager-879f6c89f-czl94\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.560076 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-etcd-client\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.577436 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa151fb7-08d1-48b4-b477-e3310ff2274d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2trtv\" (UID: \"aa151fb7-08d1-48b4-b477-e3310ff2274d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.577445 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1cc84826-c569-408e-94e9-8cf9bc49eae9-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-lcfx6\" (UID: \"1cc84826-c569-408e-94e9-8cf9bc49eae9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.578101 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a0dc011-d040-4ec5-babf-d95b03ec699e-config\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.578154 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.589079 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-client-ca\") pod \"controller-manager-879f6c89f-czl94\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.589422 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.589426 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a0dc011-d040-4ec5-babf-d95b03ec699e-serving-cert\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.593975 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa151fb7-08d1-48b4-b477-e3310ff2274d-serving-cert\") pod \"authentication-operator-69f744f599-2trtv\" (UID: \"aa151fb7-08d1-48b4-b477-e3310ff2274d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.594392 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e51eada8-802d-4064-8705-2647ffa44189-serving-cert\") pod \"controller-manager-879f6c89f-czl94\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.594571 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-encryption-config\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.594829 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-serving-cert\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.594937 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.595187 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80cf56c0-7055-4e98-9fad-f728dd8b992a-serving-cert\") pod \"route-controller-manager-6576b87f9c-plqjj\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.595620 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9a0dc011-d040-4ec5-babf-d95b03ec699e-etcd-client\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.595701 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9a0dc011-d040-4ec5-babf-d95b03ec699e-encryption-config\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.604659 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8936a9d-f33c-4de3-a759-7df7cf3e48b2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-g2szh\" (UID: \"c8936a9d-f33c-4de3-a759-7df7cf3e48b2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.607445 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hqnkq"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.608005 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f653a1d6-4938-418c-9b40-ad23d437a156-machine-approver-tls\") pod \"machine-approver-56656f9798-p7xvl\" (UID: \"f653a1d6-4938-418c-9b40-ad23d437a156\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.608674 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.610253 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.618698 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.622372 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.639025 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.639060 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.639070 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6rqmm"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.639081 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.639091 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.639897 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.641094 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.659791 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k68fq\" (UniqueName: \"kubernetes.io/projected/53824d3c-1dd1-4f25-b699-de9230490c84-kube-api-access-k68fq\") pod \"openshift-config-operator-7777fb866f-697vl\" (UID: \"53824d3c-1dd1-4f25-b699-de9230490c84\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.659881 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53824d3c-1dd1-4f25-b699-de9230490c84-serving-cert\") pod \"openshift-config-operator-7777fb866f-697vl\" (UID: \"53824d3c-1dd1-4f25-b699-de9230490c84\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.659905 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/53824d3c-1dd1-4f25-b699-de9230490c84-available-featuregates\") pod \"openshift-config-operator-7777fb866f-697vl\" (UID: \"53824d3c-1dd1-4f25-b699-de9230490c84\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.660369 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/53824d3c-1dd1-4f25-b699-de9230490c84-available-featuregates\") pod \"openshift-config-operator-7777fb866f-697vl\" (UID: \"53824d3c-1dd1-4f25-b699-de9230490c84\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.667494 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.672105 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53824d3c-1dd1-4f25-b699-de9230490c84-serving-cert\") pod \"openshift-config-operator-7777fb866f-697vl\" (UID: \"53824d3c-1dd1-4f25-b699-de9230490c84\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.676256 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.679879 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.682976 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.685573 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.686764 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-prp8k"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.687213 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-vtndh"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.688213 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.689825 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-55nsv"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.690344 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mbfc8"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.691147 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-4ftgs"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.692122 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-hc78d"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.692768 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hc78d" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.693668 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-czsqx"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.694719 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.694753 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-czsqx"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.695812 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hc78d"] Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.704977 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.725182 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.745396 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.765144 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.785199 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.805233 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.825628 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.846463 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.866075 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.885582 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.905374 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.925169 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.945083 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.966581 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 10:55:57 crc kubenswrapper[4646]: I1203 10:55:57.985818 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.007441 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.026149 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.054595 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.065545 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.086309 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.105991 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.125395 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.145790 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.165776 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.186412 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.205655 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.225742 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.246232 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.266551 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.285369 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.305841 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.325818 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.346761 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.366739 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.386175 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.405579 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.426396 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.446928 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.466318 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.486899 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.503734 4646 request.go:700] Waited for 1.001904635s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dolm-operator-serving-cert&limit=500&resourceVersion=0 Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.505961 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.526330 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.547034 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.566642 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.586468 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.606208 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.626253 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.646755 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.666853 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.686558 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.706870 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.726008 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.746819 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.766113 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.785490 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.805931 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.826402 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.845776 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.865493 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.888288 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.905355 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.925688 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.946164 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.966289 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 10:55:58 crc kubenswrapper[4646]: I1203 10:55:58.985076 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.006758 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.025246 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.045454 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.065408 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.085415 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.106522 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.133796 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.148303 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.165782 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.186146 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.206486 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.225604 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.246996 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.265867 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.328659 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wnzv\" (UniqueName: \"kubernetes.io/projected/e51eada8-802d-4064-8705-2647ffa44189-kube-api-access-8wnzv\") pod \"controller-manager-879f6c89f-czl94\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.350855 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68dk6\" (UniqueName: \"kubernetes.io/projected/c8936a9d-f33c-4de3-a759-7df7cf3e48b2-kube-api-access-68dk6\") pod \"openshift-apiserver-operator-796bbdcf4f-g2szh\" (UID: \"c8936a9d-f33c-4de3-a759-7df7cf3e48b2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.370308 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d722\" (UniqueName: \"kubernetes.io/projected/aa151fb7-08d1-48b4-b477-e3310ff2274d-kube-api-access-8d722\") pod \"authentication-operator-69f744f599-2trtv\" (UID: \"aa151fb7-08d1-48b4-b477-e3310ff2274d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.386037 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blsgz\" (UniqueName: \"kubernetes.io/projected/c202a4cb-ebe8-419d-a42d-27a286eeb4a0-kube-api-access-blsgz\") pod \"apiserver-7bbb656c7d-j6mst\" (UID: \"c202a4cb-ebe8-419d-a42d-27a286eeb4a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.417093 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn2fx\" (UniqueName: \"kubernetes.io/projected/f653a1d6-4938-418c-9b40-ad23d437a156-kube-api-access-hn2fx\") pod \"machine-approver-56656f9798-p7xvl\" (UID: \"f653a1d6-4938-418c-9b40-ad23d437a156\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.425552 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dq2p\" (UniqueName: \"kubernetes.io/projected/9a0dc011-d040-4ec5-babf-d95b03ec699e-kube-api-access-2dq2p\") pod \"apiserver-76f77b778f-q7bzw\" (UID: \"9a0dc011-d040-4ec5-babf-d95b03ec699e\") " pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.448799 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.451630 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfjx4\" (UniqueName: \"kubernetes.io/projected/1cc84826-c569-408e-94e9-8cf9bc49eae9-kube-api-access-cfjx4\") pod \"machine-api-operator-5694c8668f-lcfx6\" (UID: \"1cc84826-c569-408e-94e9-8cf9bc49eae9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.473619 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.480562 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-578ds\" (UniqueName: \"kubernetes.io/projected/80cf56c0-7055-4e98-9fad-f728dd8b992a-kube-api-access-578ds\") pod \"route-controller-manager-6576b87f9c-plqjj\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.494108 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.497475 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k68fq\" (UniqueName: \"kubernetes.io/projected/53824d3c-1dd1-4f25-b699-de9230490c84-kube-api-access-k68fq\") pod \"openshift-config-operator-7777fb866f-697vl\" (UID: \"53824d3c-1dd1-4f25-b699-de9230490c84\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.504282 4646 request.go:700] Waited for 1.811235392s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/configmaps?fieldSelector=metadata.name%3Ddns-default&limit=500&resourceVersion=0 Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.507017 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.525880 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.526257 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.547072 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.549685 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.565674 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.583946 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm9r5\" (UniqueName: \"kubernetes.io/projected/341c245b-3af9-42fd-9592-00737bcc7ba7-kube-api-access-rm9r5\") pod \"kube-storage-version-migrator-operator-b67b599dd-cb8tj\" (UID: \"341c245b-3af9-42fd-9592-00737bcc7ba7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.584080 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.584132 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw9p8\" (UniqueName: \"kubernetes.io/projected/500b7283-f949-4ff4-a868-024bc3a8a6e5-kube-api-access-qw9p8\") pod \"cluster-image-registry-operator-dc59b4c8b-wndxj\" (UID: \"500b7283-f949-4ff4-a868-024bc3a8a6e5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.584185 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c656cab0-e64d-4a92-96d5-3239ec762208-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-ztz2p\" (UID: \"c656cab0-e64d-4a92-96d5-3239ec762208\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.584231 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jcw4\" (UniqueName: \"kubernetes.io/projected/54c0778a-7e52-4da6-ab86-6a3492eb278b-kube-api-access-6jcw4\") pod \"downloads-7954f5f757-l2kck\" (UID: \"54c0778a-7e52-4da6-ab86-6a3492eb278b\") " pod="openshift-console/downloads-7954f5f757-l2kck" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.584497 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4qzw\" (UniqueName: \"kubernetes.io/projected/e6da1553-a91a-4055-af3d-bd30b20e6c0f-kube-api-access-x4qzw\") pod \"console-operator-58897d9998-52ndk\" (UID: \"e6da1553-a91a-4055-af3d-bd30b20e6c0f\") " pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.585540 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-registry-tls\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.585620 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-bound-sa-token\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.585672 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.585725 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.585806 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-trusted-ca-bundle\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.585900 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.586014 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-audit-dir\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.586058 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-config\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.586155 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6da1553-a91a-4055-af3d-bd30b20e6c0f-serving-cert\") pod \"console-operator-58897d9998-52ndk\" (UID: \"e6da1553-a91a-4055-af3d-bd30b20e6c0f\") " pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.586197 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c656cab0-e64d-4a92-96d5-3239ec762208-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-ztz2p\" (UID: \"c656cab0-e64d-4a92-96d5-3239ec762208\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.586270 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/500b7283-f949-4ff4-a868-024bc3a8a6e5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-wndxj\" (UID: \"500b7283-f949-4ff4-a868-024bc3a8a6e5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.586373 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-trusted-ca\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.586415 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/61e9e161-fb19-4da8-9e38-4ca880134e94-metrics-tls\") pod \"dns-operator-744455d44c-4jgxv\" (UID: \"61e9e161-fb19-4da8-9e38-4ca880134e94\") " pod="openshift-dns-operator/dns-operator-744455d44c-4jgxv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.586469 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-oauth-serving-cert\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.586659 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-registry-certificates\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.586704 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/500b7283-f949-4ff4-a868-024bc3a8a6e5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-wndxj\" (UID: \"500b7283-f949-4ff4-a868-024bc3a8a6e5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.586762 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-serving-cert\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.586795 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-oauth-config\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.586827 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xz4r\" (UniqueName: \"kubernetes.io/projected/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-kube-api-access-2xz4r\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.586863 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/500b7283-f949-4ff4-a868-024bc3a8a6e5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-wndxj\" (UID: \"500b7283-f949-4ff4-a868-024bc3a8a6e5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.586919 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587075 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78ptz\" (UniqueName: \"kubernetes.io/projected/61e9e161-fb19-4da8-9e38-4ca880134e94-kube-api-access-78ptz\") pod \"dns-operator-744455d44c-4jgxv\" (UID: \"61e9e161-fb19-4da8-9e38-4ca880134e94\") " pod="openshift-dns-operator/dns-operator-744455d44c-4jgxv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587122 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/341c245b-3af9-42fd-9592-00737bcc7ba7-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-cb8tj\" (UID: \"341c245b-3af9-42fd-9592-00737bcc7ba7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587178 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-ca-trust-extracted\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587226 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/341c245b-3af9-42fd-9592-00737bcc7ba7-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-cb8tj\" (UID: \"341c245b-3af9-42fd-9592-00737bcc7ba7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587261 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587320 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-service-ca\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587398 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587449 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587500 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f20ff0d2-d132-4a30-83b5-e59f6e7260bf-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-sc5lh\" (UID: \"f20ff0d2-d132-4a30-83b5-e59f6e7260bf\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sc5lh" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587533 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9jsx\" (UniqueName: \"kubernetes.io/projected/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-kube-api-access-m9jsx\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587588 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587631 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587666 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e6da1553-a91a-4055-af3d-bd30b20e6c0f-trusted-ca\") pod \"console-operator-58897d9998-52ndk\" (UID: \"e6da1553-a91a-4055-af3d-bd30b20e6c0f\") " pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587794 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx6hp\" (UniqueName: \"kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-kube-api-access-tx6hp\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587838 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfvbk\" (UniqueName: \"kubernetes.io/projected/f20ff0d2-d132-4a30-83b5-e59f6e7260bf-kube-api-access-wfvbk\") pod \"cluster-samples-operator-665b6dd947-sc5lh\" (UID: \"f20ff0d2-d132-4a30-83b5-e59f6e7260bf\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sc5lh" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587873 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587916 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq8vl\" (UniqueName: \"kubernetes.io/projected/c656cab0-e64d-4a92-96d5-3239ec762208-kube-api-access-hq8vl\") pod \"openshift-controller-manager-operator-756b6f6bc6-ztz2p\" (UID: \"c656cab0-e64d-4a92-96d5-3239ec762208\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.587988 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-installation-pull-secrets\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.588044 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-audit-policies\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.588082 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.588119 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6da1553-a91a-4055-af3d-bd30b20e6c0f-config\") pod \"console-operator-58897d9998-52ndk\" (UID: \"e6da1553-a91a-4055-af3d-bd30b20e6c0f\") " pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.594217 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 10:55:59 crc kubenswrapper[4646]: E1203 10:55:59.598666 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:00.098652278 +0000 UTC m=+136.561708413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.599446 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.600772 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.612612 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.615472 4646 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.644180 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" event={"ID":"f653a1d6-4938-418c-9b40-ad23d437a156","Type":"ContainerStarted","Data":"a3c4afd5ba3209d063e38e5cc6cf099dbb842824074e7cb269adc04e7103ea0a"} Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.682621 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.690795 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697428 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-trusted-ca-bundle\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697494 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lkssm\" (UID: \"9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697541 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697563 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-config-volume\") pod \"collect-profiles-29412645-6w78t\" (UID: \"d8e77566-f3a2-4a10-84b5-4e6078bc6f23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697587 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-config\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697607 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/af7f7f9c-e126-413f-8d2d-499af6740741-profile-collector-cert\") pod \"catalog-operator-68c6474976-phkk4\" (UID: \"af7f7f9c-e126-413f-8d2d-499af6740741\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697629 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9dc04fc6-6a79-4317-8b14-be963345a26c-proxy-tls\") pod \"machine-config-controller-84d6567774-cg4cm\" (UID: \"9dc04fc6-6a79-4317-8b14-be963345a26c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697649 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6shw\" (UniqueName: \"kubernetes.io/projected/578c8a44-77e3-4c21-b841-3fc93a82b0f1-kube-api-access-g6shw\") pod \"package-server-manager-789f6589d5-2vh6v\" (UID: \"578c8a44-77e3-4c21-b841-3fc93a82b0f1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697703 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c656cab0-e64d-4a92-96d5-3239ec762208-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-ztz2p\" (UID: \"c656cab0-e64d-4a92-96d5-3239ec762208\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697729 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/500b7283-f949-4ff4-a868-024bc3a8a6e5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-wndxj\" (UID: \"500b7283-f949-4ff4-a868-024bc3a8a6e5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697750 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cc1272d-e287-4e28-92f7-27dc0aa0d927-config\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697771 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-secret-volume\") pod \"collect-profiles-29412645-6w78t\" (UID: \"d8e77566-f3a2-4a10-84b5-4e6078bc6f23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697794 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9ztb\" (UniqueName: \"kubernetes.io/projected/75e92964-1833-4889-b175-16f4bba0ed1b-kube-api-access-q9ztb\") pod \"migrator-59844c95c7-prp8k\" (UID: \"75e92964-1833-4889-b175-16f4bba0ed1b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-prp8k" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697831 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cc1272d-e287-4e28-92f7-27dc0aa0d927-serving-cert\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697853 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cd573c9a-19ea-4975-9645-9d5b9e258d58-tmpfs\") pod \"packageserver-d55dfcdfc-9x4fj\" (UID: \"cd573c9a-19ea-4975-9645-9d5b9e258d58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697878 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78ptz\" (UniqueName: \"kubernetes.io/projected/61e9e161-fb19-4da8-9e38-4ca880134e94-kube-api-access-78ptz\") pod \"dns-operator-744455d44c-4jgxv\" (UID: \"61e9e161-fb19-4da8-9e38-4ca880134e94\") " pod="openshift-dns-operator/dns-operator-744455d44c-4jgxv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697897 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dswf8\" (UniqueName: \"kubernetes.io/projected/af7f7f9c-e126-413f-8d2d-499af6740741-kube-api-access-dswf8\") pod \"catalog-operator-68c6474976-phkk4\" (UID: \"af7f7f9c-e126-413f-8d2d-499af6740741\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697918 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/45f218e8-2c99-412f-b85e-956417b5d74a-signing-key\") pod \"service-ca-9c57cc56f-6rqmm\" (UID: \"45f218e8-2c99-412f-b85e-956417b5d74a\") " pod="openshift-service-ca/service-ca-9c57cc56f-6rqmm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697936 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bhzv\" (UniqueName: \"kubernetes.io/projected/92576279-7206-4601-9734-85243e90b5ed-kube-api-access-8bhzv\") pod \"dns-default-hc78d\" (UID: \"92576279-7206-4601-9734-85243e90b5ed\") " pod="openshift-dns/dns-default-hc78d" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697956 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjwhf\" (UniqueName: \"kubernetes.io/projected/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-kube-api-access-tjwhf\") pod \"collect-profiles-29412645-6w78t\" (UID: \"d8e77566-f3a2-4a10-84b5-4e6078bc6f23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697976 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/92576279-7206-4601-9734-85243e90b5ed-metrics-tls\") pod \"dns-default-hc78d\" (UID: \"92576279-7206-4601-9734-85243e90b5ed\") " pod="openshift-dns/dns-default-hc78d" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.697999 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/341c245b-3af9-42fd-9592-00737bcc7ba7-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-cb8tj\" (UID: \"341c245b-3af9-42fd-9592-00737bcc7ba7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698016 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/324debd2-da21-490a-9e83-36fbcb470aab-images\") pod \"machine-config-operator-74547568cd-ntvnm\" (UID: \"324debd2-da21-490a-9e83-36fbcb470aab\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698038 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4f9n\" (UniqueName: \"kubernetes.io/projected/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-kube-api-access-m4f9n\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698057 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/af7f7f9c-e126-413f-8d2d-499af6740741-srv-cert\") pod \"catalog-operator-68c6474976-phkk4\" (UID: \"af7f7f9c-e126-413f-8d2d-499af6740741\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698076 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwtbr\" (UniqueName: \"kubernetes.io/projected/3cc1272d-e287-4e28-92f7-27dc0aa0d927-kube-api-access-cwtbr\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698096 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c5ac722-cd40-44e9-921b-968dbe5e33f1-config\") pod \"service-ca-operator-777779d784-wgpqg\" (UID: \"4c5ac722-cd40-44e9-921b-968dbe5e33f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698134 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698154 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9jsx\" (UniqueName: \"kubernetes.io/projected/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-kube-api-access-m9jsx\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698172 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698191 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698211 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5wscd\" (UID: \"697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698235 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e6da1553-a91a-4055-af3d-bd30b20e6c0f-trusted-ca\") pod \"console-operator-58897d9998-52ndk\" (UID: \"e6da1553-a91a-4055-af3d-bd30b20e6c0f\") " pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698268 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbjz4\" (UniqueName: \"kubernetes.io/projected/9dc04fc6-6a79-4317-8b14-be963345a26c-kube-api-access-mbjz4\") pod \"machine-config-controller-84d6567774-cg4cm\" (UID: \"9dc04fc6-6a79-4317-8b14-be963345a26c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698287 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-socket-dir\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698310 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx6hp\" (UniqueName: \"kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-kube-api-access-tx6hp\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698367 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfvbk\" (UniqueName: \"kubernetes.io/projected/f20ff0d2-d132-4a30-83b5-e59f6e7260bf-kube-api-access-wfvbk\") pod \"cluster-samples-operator-665b6dd947-sc5lh\" (UID: \"f20ff0d2-d132-4a30-83b5-e59f6e7260bf\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sc5lh" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698389 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/464c6029-4dac-48ee-9f4f-70659bb6445b-config\") pod \"kube-apiserver-operator-766d6c64bb-zmnzd\" (UID: \"464c6029-4dac-48ee-9f4f-70659bb6445b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698410 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/143daef8-3687-4018-8f9f-2a675eac0909-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mbfc8\" (UID: \"143daef8-3687-4018-8f9f-2a675eac0909\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698433 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-audit-policies\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698458 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698474 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6da1553-a91a-4055-af3d-bd30b20e6c0f-config\") pod \"console-operator-58897d9998-52ndk\" (UID: \"e6da1553-a91a-4055-af3d-bd30b20e6c0f\") " pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698496 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-csi-data-dir\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698526 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/143daef8-3687-4018-8f9f-2a675eac0909-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mbfc8\" (UID: \"143daef8-3687-4018-8f9f-2a675eac0909\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698555 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698589 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw9p8\" (UniqueName: \"kubernetes.io/projected/500b7283-f949-4ff4-a868-024bc3a8a6e5-kube-api-access-qw9p8\") pod \"cluster-image-registry-operator-dc59b4c8b-wndxj\" (UID: \"500b7283-f949-4ff4-a868-024bc3a8a6e5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698628 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9p4v\" (UniqueName: \"kubernetes.io/projected/cd573c9a-19ea-4975-9645-9d5b9e258d58-kube-api-access-f9p4v\") pod \"packageserver-d55dfcdfc-9x4fj\" (UID: \"cd573c9a-19ea-4975-9645-9d5b9e258d58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698667 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xsm7\" (UniqueName: \"kubernetes.io/projected/88293fb3-626a-4944-9d1f-2c569832661d-kube-api-access-4xsm7\") pod \"olm-operator-6b444d44fb-pkrnf\" (UID: \"88293fb3-626a-4944-9d1f-2c569832661d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698705 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c656cab0-e64d-4a92-96d5-3239ec762208-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-ztz2p\" (UID: \"c656cab0-e64d-4a92-96d5-3239ec762208\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698741 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jcw4\" (UniqueName: \"kubernetes.io/projected/54c0778a-7e52-4da6-ab86-6a3492eb278b-kube-api-access-6jcw4\") pod \"downloads-7954f5f757-l2kck\" (UID: \"54c0778a-7e52-4da6-ab86-6a3492eb278b\") " pod="openshift-console/downloads-7954f5f757-l2kck" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698763 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9dc04fc6-6a79-4317-8b14-be963345a26c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-cg4cm\" (UID: \"9dc04fc6-6a79-4317-8b14-be963345a26c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698806 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/324debd2-da21-490a-9e83-36fbcb470aab-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ntvnm\" (UID: \"324debd2-da21-490a-9e83-36fbcb470aab\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698842 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4qzw\" (UniqueName: \"kubernetes.io/projected/e6da1553-a91a-4055-af3d-bd30b20e6c0f-kube-api-access-x4qzw\") pod \"console-operator-58897d9998-52ndk\" (UID: \"e6da1553-a91a-4055-af3d-bd30b20e6c0f\") " pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698879 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rghjj\" (UniqueName: \"kubernetes.io/projected/45f218e8-2c99-412f-b85e-956417b5d74a-kube-api-access-rghjj\") pod \"service-ca-9c57cc56f-6rqmm\" (UID: \"45f218e8-2c99-412f-b85e-956417b5d74a\") " pod="openshift-service-ca/service-ca-9c57cc56f-6rqmm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698948 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-registry-tls\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.698987 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-bound-sa-token\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699068 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699090 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa99705d-9e2f-4c32-85cb-04f0cde27411-service-ca-bundle\") pod \"router-default-5444994796-7whtv\" (UID: \"aa99705d-9e2f-4c32-85cb-04f0cde27411\") " pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699113 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/324debd2-da21-490a-9e83-36fbcb470aab-proxy-tls\") pod \"machine-config-operator-74547568cd-ntvnm\" (UID: \"324debd2-da21-490a-9e83-36fbcb470aab\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699131 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f9912d55-27ed-44cf-a597-9591fa35476c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hqnkq\" (UID: \"f9912d55-27ed-44cf-a597-9591fa35476c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hqnkq" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699154 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-audit-dir\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699174 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6da1553-a91a-4055-af3d-bd30b20e6c0f-serving-cert\") pod \"console-operator-58897d9998-52ndk\" (UID: \"e6da1553-a91a-4055-af3d-bd30b20e6c0f\") " pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699194 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77pxz\" (UniqueName: \"kubernetes.io/projected/40507d60-aae3-43da-85b0-09c02327ece8-kube-api-access-77pxz\") pod \"ingress-canary-vtndh\" (UID: \"40507d60-aae3-43da-85b0-09c02327ece8\") " pod="openshift-ingress-canary/ingress-canary-vtndh" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699209 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/aa99705d-9e2f-4c32-85cb-04f0cde27411-default-certificate\") pod \"router-default-5444994796-7whtv\" (UID: \"aa99705d-9e2f-4c32-85cb-04f0cde27411\") " pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699228 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/464c6029-4dac-48ee-9f4f-70659bb6445b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zmnzd\" (UID: \"464c6029-4dac-48ee-9f4f-70659bb6445b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699247 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz558\" (UniqueName: \"kubernetes.io/projected/f9912d55-27ed-44cf-a597-9591fa35476c-kube-api-access-tz558\") pod \"multus-admission-controller-857f4d67dd-hqnkq\" (UID: \"f9912d55-27ed-44cf-a597-9591fa35476c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hqnkq" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699265 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4b507901-6faf-4fcf-a512-e876b56c54e1-node-bootstrap-token\") pod \"machine-config-server-mvqf6\" (UID: \"4b507901-6faf-4fcf-a512-e876b56c54e1\") " pod="openshift-machine-config-operator/machine-config-server-mvqf6" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699283 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450-trusted-ca\") pod \"ingress-operator-5b745b69d9-lkssm\" (UID: \"9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699303 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-trusted-ca\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699323 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/61e9e161-fb19-4da8-9e38-4ca880134e94-metrics-tls\") pod \"dns-operator-744455d44c-4jgxv\" (UID: \"61e9e161-fb19-4da8-9e38-4ca880134e94\") " pod="openshift-dns-operator/dns-operator-744455d44c-4jgxv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699355 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aa99705d-9e2f-4c32-85cb-04f0cde27411-metrics-certs\") pod \"router-default-5444994796-7whtv\" (UID: \"aa99705d-9e2f-4c32-85cb-04f0cde27411\") " pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699373 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/88293fb3-626a-4944-9d1f-2c569832661d-srv-cert\") pod \"olm-operator-6b444d44fb-pkrnf\" (UID: \"88293fb3-626a-4944-9d1f-2c569832661d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699422 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-oauth-serving-cert\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699479 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5wscd\" (UID: \"697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699497 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/aa99705d-9e2f-4c32-85cb-04f0cde27411-stats-auth\") pod \"router-default-5444994796-7whtv\" (UID: \"aa99705d-9e2f-4c32-85cb-04f0cde27411\") " pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699513 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3cc1272d-e287-4e28-92f7-27dc0aa0d927-etcd-service-ca\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699532 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-registry-certificates\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699550 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwsqc\" (UniqueName: \"kubernetes.io/projected/4b507901-6faf-4fcf-a512-e876b56c54e1-kube-api-access-kwsqc\") pod \"machine-config-server-mvqf6\" (UID: \"4b507901-6faf-4fcf-a512-e876b56c54e1\") " pod="openshift-machine-config-operator/machine-config-server-mvqf6" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699669 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cd573c9a-19ea-4975-9645-9d5b9e258d58-webhook-cert\") pod \"packageserver-d55dfcdfc-9x4fj\" (UID: \"cd573c9a-19ea-4975-9645-9d5b9e258d58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699690 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-serving-cert\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699706 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-oauth-config\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699725 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xz4r\" (UniqueName: \"kubernetes.io/projected/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-kube-api-access-2xz4r\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699746 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/500b7283-f949-4ff4-a868-024bc3a8a6e5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-wndxj\" (UID: \"500b7283-f949-4ff4-a868-024bc3a8a6e5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699812 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cd573c9a-19ea-4975-9645-9d5b9e258d58-apiservice-cert\") pod \"packageserver-d55dfcdfc-9x4fj\" (UID: \"cd573c9a-19ea-4975-9645-9d5b9e258d58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699842 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/500b7283-f949-4ff4-a868-024bc3a8a6e5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-wndxj\" (UID: \"500b7283-f949-4ff4-a868-024bc3a8a6e5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699872 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/341c245b-3af9-42fd-9592-00737bcc7ba7-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-cb8tj\" (UID: \"341c245b-3af9-42fd-9592-00737bcc7ba7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699904 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-ca-trust-extracted\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.699982 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-registration-dir\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.700002 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/88293fb3-626a-4944-9d1f-2c569832661d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pkrnf\" (UID: \"88293fb3-626a-4944-9d1f-2c569832661d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.700036 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/578c8a44-77e3-4c21-b841-3fc93a82b0f1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2vh6v\" (UID: \"578c8a44-77e3-4c21-b841-3fc93a82b0f1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.700057 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/682f390b-3453-40c5-aa8c-5a834ca1d599-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5dz9v\" (UID: \"682f390b-3453-40c5-aa8c-5a834ca1d599\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.700073 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5wscd\" (UID: \"697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.700135 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4b507901-6faf-4fcf-a512-e876b56c54e1-certs\") pod \"machine-config-server-mvqf6\" (UID: \"4b507901-6faf-4fcf-a512-e876b56c54e1\") " pod="openshift-machine-config-operator/machine-config-server-mvqf6" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.700148 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-trusted-ca-bundle\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.700171 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.700194 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmdqv\" (UniqueName: \"kubernetes.io/projected/aa99705d-9e2f-4c32-85cb-04f0cde27411-kube-api-access-kmdqv\") pod \"router-default-5444994796-7whtv\" (UID: \"aa99705d-9e2f-4c32-85cb-04f0cde27411\") " pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.700216 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-service-ca\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.700278 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/682f390b-3453-40c5-aa8c-5a834ca1d599-config\") pod \"kube-controller-manager-operator-78b949d7b-5dz9v\" (UID: \"682f390b-3453-40c5-aa8c-5a834ca1d599\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.700309 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.701150 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/341c245b-3af9-42fd-9592-00737bcc7ba7-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-cb8tj\" (UID: \"341c245b-3af9-42fd-9592-00737bcc7ba7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.701622 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.702165 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-config\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.705212 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: E1203 10:55:59.705980 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:00.205963532 +0000 UTC m=+136.669019667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.706112 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.708543 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.708827 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e6da1553-a91a-4055-af3d-bd30b20e6c0f-trusted-ca\") pod \"console-operator-58897d9998-52ndk\" (UID: \"e6da1553-a91a-4055-af3d-bd30b20e6c0f\") " pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.709502 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c656cab0-e64d-4a92-96d5-3239ec762208-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-ztz2p\" (UID: \"c656cab0-e64d-4a92-96d5-3239ec762208\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710034 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40507d60-aae3-43da-85b0-09c02327ece8-cert\") pod \"ingress-canary-vtndh\" (UID: \"40507d60-aae3-43da-85b0-09c02327ece8\") " pod="openshift-ingress-canary/ingress-canary-vtndh" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710082 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-mountpoint-dir\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710107 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/45f218e8-2c99-412f-b85e-956417b5d74a-signing-cabundle\") pod \"service-ca-9c57cc56f-6rqmm\" (UID: \"45f218e8-2c99-412f-b85e-956417b5d74a\") " pod="openshift-service-ca/service-ca-9c57cc56f-6rqmm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710199 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f20ff0d2-d132-4a30-83b5-e59f6e7260bf-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-sc5lh\" (UID: \"f20ff0d2-d132-4a30-83b5-e59f6e7260bf\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sc5lh" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710222 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/682f390b-3453-40c5-aa8c-5a834ca1d599-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5dz9v\" (UID: \"682f390b-3453-40c5-aa8c-5a834ca1d599\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710243 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c5ac722-cd40-44e9-921b-968dbe5e33f1-serving-cert\") pod \"service-ca-operator-777779d784-wgpqg\" (UID: \"4c5ac722-cd40-44e9-921b-968dbe5e33f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710277 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-plugins-dir\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710371 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92576279-7206-4601-9734-85243e90b5ed-config-volume\") pod \"dns-default-hc78d\" (UID: \"92576279-7206-4601-9734-85243e90b5ed\") " pod="openshift-dns/dns-default-hc78d" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710395 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wwc5\" (UniqueName: \"kubernetes.io/projected/4c5ac722-cd40-44e9-921b-968dbe5e33f1-kube-api-access-2wwc5\") pod \"service-ca-operator-777779d784-wgpqg\" (UID: \"4c5ac722-cd40-44e9-921b-968dbe5e33f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710427 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710451 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq8vl\" (UniqueName: \"kubernetes.io/projected/c656cab0-e64d-4a92-96d5-3239ec762208-kube-api-access-hq8vl\") pod \"openshift-controller-manager-operator-756b6f6bc6-ztz2p\" (UID: \"c656cab0-e64d-4a92-96d5-3239ec762208\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710468 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3cc1272d-e287-4e28-92f7-27dc0aa0d927-etcd-client\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710548 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghzsx\" (UniqueName: \"kubernetes.io/projected/143daef8-3687-4018-8f9f-2a675eac0909-kube-api-access-ghzsx\") pod \"marketplace-operator-79b997595-mbfc8\" (UID: \"143daef8-3687-4018-8f9f-2a675eac0909\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710572 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-installation-pull-secrets\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710593 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xknkv\" (UniqueName: \"kubernetes.io/projected/324debd2-da21-490a-9e83-36fbcb470aab-kube-api-access-xknkv\") pod \"machine-config-operator-74547568cd-ntvnm\" (UID: \"324debd2-da21-490a-9e83-36fbcb470aab\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710614 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm9r5\" (UniqueName: \"kubernetes.io/projected/341c245b-3af9-42fd-9592-00737bcc7ba7-kube-api-access-rm9r5\") pod \"kube-storage-version-migrator-operator-b67b599dd-cb8tj\" (UID: \"341c245b-3af9-42fd-9592-00737bcc7ba7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710635 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76dnx\" (UniqueName: \"kubernetes.io/projected/9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450-kube-api-access-76dnx\") pod \"ingress-operator-5b745b69d9-lkssm\" (UID: \"9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710727 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjkt4\" (UniqueName: \"kubernetes.io/projected/041d1b1a-4861-4b7e-98ee-f5707404c173-kube-api-access-sjkt4\") pod \"control-plane-machine-set-operator-78cbb6b69f-55nsv\" (UID: \"041d1b1a-4861-4b7e-98ee-f5707404c173\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-55nsv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710749 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3cc1272d-e287-4e28-92f7-27dc0aa0d927-etcd-ca\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710785 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/041d1b1a-4861-4b7e-98ee-f5707404c173-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-55nsv\" (UID: \"041d1b1a-4861-4b7e-98ee-f5707404c173\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-55nsv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710805 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450-metrics-tls\") pod \"ingress-operator-5b745b69d9-lkssm\" (UID: \"9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710872 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/464c6029-4dac-48ee-9f4f-70659bb6445b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zmnzd\" (UID: \"464c6029-4dac-48ee-9f4f-70659bb6445b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.710901 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.711647 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-audit-policies\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.716778 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-service-ca\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.718050 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-oauth-serving-cert\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.718972 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/500b7283-f949-4ff4-a868-024bc3a8a6e5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-wndxj\" (UID: \"500b7283-f949-4ff4-a868-024bc3a8a6e5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.720144 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-registry-certificates\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.723053 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.723250 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-audit-dir\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.725384 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6da1553-a91a-4055-af3d-bd30b20e6c0f-config\") pod \"console-operator-58897d9998-52ndk\" (UID: \"e6da1553-a91a-4055-af3d-bd30b20e6c0f\") " pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.725793 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c656cab0-e64d-4a92-96d5-3239ec762208-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-ztz2p\" (UID: \"c656cab0-e64d-4a92-96d5-3239ec762208\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.735061 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-ca-trust-extracted\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.736742 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-serving-cert\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.737556 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/500b7283-f949-4ff4-a868-024bc3a8a6e5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-wndxj\" (UID: \"500b7283-f949-4ff4-a868-024bc3a8a6e5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.738088 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-trusted-ca\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.738620 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.750380 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.750441 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/341c245b-3af9-42fd-9592-00737bcc7ba7-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-cb8tj\" (UID: \"341c245b-3af9-42fd-9592-00737bcc7ba7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.750910 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6da1553-a91a-4055-af3d-bd30b20e6c0f-serving-cert\") pod \"console-operator-58897d9998-52ndk\" (UID: \"e6da1553-a91a-4055-af3d-bd30b20e6c0f\") " pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.752202 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.752451 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-installation-pull-secrets\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.755246 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.756087 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst"] Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.756232 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.762281 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/61e9e161-fb19-4da8-9e38-4ca880134e94-metrics-tls\") pod \"dns-operator-744455d44c-4jgxv\" (UID: \"61e9e161-fb19-4da8-9e38-4ca880134e94\") " pod="openshift-dns-operator/dns-operator-744455d44c-4jgxv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.765218 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.766988 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-registry-tls\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.769975 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-oauth-config\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.787692 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9jsx\" (UniqueName: \"kubernetes.io/projected/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-kube-api-access-m9jsx\") pod \"oauth-openshift-558db77b4-nb6wv\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.788114 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78ptz\" (UniqueName: \"kubernetes.io/projected/61e9e161-fb19-4da8-9e38-4ca880134e94-kube-api-access-78ptz\") pod \"dns-operator-744455d44c-4jgxv\" (UID: \"61e9e161-fb19-4da8-9e38-4ca880134e94\") " pod="openshift-dns-operator/dns-operator-744455d44c-4jgxv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.788361 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f20ff0d2-d132-4a30-83b5-e59f6e7260bf-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-sc5lh\" (UID: \"f20ff0d2-d132-4a30-83b5-e59f6e7260bf\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sc5lh" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.804692 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx6hp\" (UniqueName: \"kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-kube-api-access-tx6hp\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.808103 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-4jgxv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.811683 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-czl94"] Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812131 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dswf8\" (UniqueName: \"kubernetes.io/projected/af7f7f9c-e126-413f-8d2d-499af6740741-kube-api-access-dswf8\") pod \"catalog-operator-68c6474976-phkk4\" (UID: \"af7f7f9c-e126-413f-8d2d-499af6740741\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812157 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/45f218e8-2c99-412f-b85e-956417b5d74a-signing-key\") pod \"service-ca-9c57cc56f-6rqmm\" (UID: \"45f218e8-2c99-412f-b85e-956417b5d74a\") " pod="openshift-service-ca/service-ca-9c57cc56f-6rqmm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812180 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bhzv\" (UniqueName: \"kubernetes.io/projected/92576279-7206-4601-9734-85243e90b5ed-kube-api-access-8bhzv\") pod \"dns-default-hc78d\" (UID: \"92576279-7206-4601-9734-85243e90b5ed\") " pod="openshift-dns/dns-default-hc78d" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812196 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjwhf\" (UniqueName: \"kubernetes.io/projected/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-kube-api-access-tjwhf\") pod \"collect-profiles-29412645-6w78t\" (UID: \"d8e77566-f3a2-4a10-84b5-4e6078bc6f23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812212 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/92576279-7206-4601-9734-85243e90b5ed-metrics-tls\") pod \"dns-default-hc78d\" (UID: \"92576279-7206-4601-9734-85243e90b5ed\") " pod="openshift-dns/dns-default-hc78d" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812228 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/324debd2-da21-490a-9e83-36fbcb470aab-images\") pod \"machine-config-operator-74547568cd-ntvnm\" (UID: \"324debd2-da21-490a-9e83-36fbcb470aab\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812242 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4f9n\" (UniqueName: \"kubernetes.io/projected/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-kube-api-access-m4f9n\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812256 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/af7f7f9c-e126-413f-8d2d-499af6740741-srv-cert\") pod \"catalog-operator-68c6474976-phkk4\" (UID: \"af7f7f9c-e126-413f-8d2d-499af6740741\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812270 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwtbr\" (UniqueName: \"kubernetes.io/projected/3cc1272d-e287-4e28-92f7-27dc0aa0d927-kube-api-access-cwtbr\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812285 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c5ac722-cd40-44e9-921b-968dbe5e33f1-config\") pod \"service-ca-operator-777779d784-wgpqg\" (UID: \"4c5ac722-cd40-44e9-921b-968dbe5e33f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812311 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5wscd\" (UID: \"697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812363 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbjz4\" (UniqueName: \"kubernetes.io/projected/9dc04fc6-6a79-4317-8b14-be963345a26c-kube-api-access-mbjz4\") pod \"machine-config-controller-84d6567774-cg4cm\" (UID: \"9dc04fc6-6a79-4317-8b14-be963345a26c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812388 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-socket-dir\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812415 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/464c6029-4dac-48ee-9f4f-70659bb6445b-config\") pod \"kube-apiserver-operator-766d6c64bb-zmnzd\" (UID: \"464c6029-4dac-48ee-9f4f-70659bb6445b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812431 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/143daef8-3687-4018-8f9f-2a675eac0909-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mbfc8\" (UID: \"143daef8-3687-4018-8f9f-2a675eac0909\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812446 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/143daef8-3687-4018-8f9f-2a675eac0909-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mbfc8\" (UID: \"143daef8-3687-4018-8f9f-2a675eac0909\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812460 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-csi-data-dir\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812474 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xsm7\" (UniqueName: \"kubernetes.io/projected/88293fb3-626a-4944-9d1f-2c569832661d-kube-api-access-4xsm7\") pod \"olm-operator-6b444d44fb-pkrnf\" (UID: \"88293fb3-626a-4944-9d1f-2c569832661d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.812496 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9p4v\" (UniqueName: \"kubernetes.io/projected/cd573c9a-19ea-4975-9645-9d5b9e258d58-kube-api-access-f9p4v\") pod \"packageserver-d55dfcdfc-9x4fj\" (UID: \"cd573c9a-19ea-4975-9645-9d5b9e258d58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.816239 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/143daef8-3687-4018-8f9f-2a675eac0909-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mbfc8\" (UID: \"143daef8-3687-4018-8f9f-2a675eac0909\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.816843 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9dc04fc6-6a79-4317-8b14-be963345a26c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-cg4cm\" (UID: \"9dc04fc6-6a79-4317-8b14-be963345a26c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.816989 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/324debd2-da21-490a-9e83-36fbcb470aab-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ntvnm\" (UID: \"324debd2-da21-490a-9e83-36fbcb470aab\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817040 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rghjj\" (UniqueName: \"kubernetes.io/projected/45f218e8-2c99-412f-b85e-956417b5d74a-kube-api-access-rghjj\") pod \"service-ca-9c57cc56f-6rqmm\" (UID: \"45f218e8-2c99-412f-b85e-956417b5d74a\") " pod="openshift-service-ca/service-ca-9c57cc56f-6rqmm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817061 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa99705d-9e2f-4c32-85cb-04f0cde27411-service-ca-bundle\") pod \"router-default-5444994796-7whtv\" (UID: \"aa99705d-9e2f-4c32-85cb-04f0cde27411\") " pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817088 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/324debd2-da21-490a-9e83-36fbcb470aab-proxy-tls\") pod \"machine-config-operator-74547568cd-ntvnm\" (UID: \"324debd2-da21-490a-9e83-36fbcb470aab\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817108 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f9912d55-27ed-44cf-a597-9591fa35476c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hqnkq\" (UID: \"f9912d55-27ed-44cf-a597-9591fa35476c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hqnkq" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817130 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77pxz\" (UniqueName: \"kubernetes.io/projected/40507d60-aae3-43da-85b0-09c02327ece8-kube-api-access-77pxz\") pod \"ingress-canary-vtndh\" (UID: \"40507d60-aae3-43da-85b0-09c02327ece8\") " pod="openshift-ingress-canary/ingress-canary-vtndh" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817146 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/aa99705d-9e2f-4c32-85cb-04f0cde27411-default-certificate\") pod \"router-default-5444994796-7whtv\" (UID: \"aa99705d-9e2f-4c32-85cb-04f0cde27411\") " pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817165 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/464c6029-4dac-48ee-9f4f-70659bb6445b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zmnzd\" (UID: \"464c6029-4dac-48ee-9f4f-70659bb6445b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817181 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz558\" (UniqueName: \"kubernetes.io/projected/f9912d55-27ed-44cf-a597-9591fa35476c-kube-api-access-tz558\") pod \"multus-admission-controller-857f4d67dd-hqnkq\" (UID: \"f9912d55-27ed-44cf-a597-9591fa35476c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hqnkq" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817196 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4b507901-6faf-4fcf-a512-e876b56c54e1-node-bootstrap-token\") pod \"machine-config-server-mvqf6\" (UID: \"4b507901-6faf-4fcf-a512-e876b56c54e1\") " pod="openshift-machine-config-operator/machine-config-server-mvqf6" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817220 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450-trusted-ca\") pod \"ingress-operator-5b745b69d9-lkssm\" (UID: \"9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817234 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/88293fb3-626a-4944-9d1f-2c569832661d-srv-cert\") pod \"olm-operator-6b444d44fb-pkrnf\" (UID: \"88293fb3-626a-4944-9d1f-2c569832661d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817253 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aa99705d-9e2f-4c32-85cb-04f0cde27411-metrics-certs\") pod \"router-default-5444994796-7whtv\" (UID: \"aa99705d-9e2f-4c32-85cb-04f0cde27411\") " pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817272 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3cc1272d-e287-4e28-92f7-27dc0aa0d927-etcd-service-ca\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817289 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5wscd\" (UID: \"697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817303 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/aa99705d-9e2f-4c32-85cb-04f0cde27411-stats-auth\") pod \"router-default-5444994796-7whtv\" (UID: \"aa99705d-9e2f-4c32-85cb-04f0cde27411\") " pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817320 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwsqc\" (UniqueName: \"kubernetes.io/projected/4b507901-6faf-4fcf-a512-e876b56c54e1-kube-api-access-kwsqc\") pod \"machine-config-server-mvqf6\" (UID: \"4b507901-6faf-4fcf-a512-e876b56c54e1\") " pod="openshift-machine-config-operator/machine-config-server-mvqf6" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817348 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cd573c9a-19ea-4975-9645-9d5b9e258d58-apiservice-cert\") pod \"packageserver-d55dfcdfc-9x4fj\" (UID: \"cd573c9a-19ea-4975-9645-9d5b9e258d58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817363 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cd573c9a-19ea-4975-9645-9d5b9e258d58-webhook-cert\") pod \"packageserver-d55dfcdfc-9x4fj\" (UID: \"cd573c9a-19ea-4975-9645-9d5b9e258d58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817399 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817412 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/45f218e8-2c99-412f-b85e-956417b5d74a-signing-key\") pod \"service-ca-9c57cc56f-6rqmm\" (UID: \"45f218e8-2c99-412f-b85e-956417b5d74a\") " pod="openshift-service-ca/service-ca-9c57cc56f-6rqmm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817419 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/88293fb3-626a-4944-9d1f-2c569832661d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pkrnf\" (UID: \"88293fb3-626a-4944-9d1f-2c569832661d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817493 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-registration-dir\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817524 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/578c8a44-77e3-4c21-b841-3fc93a82b0f1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2vh6v\" (UID: \"578c8a44-77e3-4c21-b841-3fc93a82b0f1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817542 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/682f390b-3453-40c5-aa8c-5a834ca1d599-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5dz9v\" (UID: \"682f390b-3453-40c5-aa8c-5a834ca1d599\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817559 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5wscd\" (UID: \"697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817574 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4b507901-6faf-4fcf-a512-e876b56c54e1-certs\") pod \"machine-config-server-mvqf6\" (UID: \"4b507901-6faf-4fcf-a512-e876b56c54e1\") " pod="openshift-machine-config-operator/machine-config-server-mvqf6" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817597 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/682f390b-3453-40c5-aa8c-5a834ca1d599-config\") pod \"kube-controller-manager-operator-78b949d7b-5dz9v\" (UID: \"682f390b-3453-40c5-aa8c-5a834ca1d599\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817619 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmdqv\" (UniqueName: \"kubernetes.io/projected/aa99705d-9e2f-4c32-85cb-04f0cde27411-kube-api-access-kmdqv\") pod \"router-default-5444994796-7whtv\" (UID: \"aa99705d-9e2f-4c32-85cb-04f0cde27411\") " pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817649 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40507d60-aae3-43da-85b0-09c02327ece8-cert\") pod \"ingress-canary-vtndh\" (UID: \"40507d60-aae3-43da-85b0-09c02327ece8\") " pod="openshift-ingress-canary/ingress-canary-vtndh" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817665 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-mountpoint-dir\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817682 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/45f218e8-2c99-412f-b85e-956417b5d74a-signing-cabundle\") pod \"service-ca-9c57cc56f-6rqmm\" (UID: \"45f218e8-2c99-412f-b85e-956417b5d74a\") " pod="openshift-service-ca/service-ca-9c57cc56f-6rqmm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817700 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/682f390b-3453-40c5-aa8c-5a834ca1d599-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5dz9v\" (UID: \"682f390b-3453-40c5-aa8c-5a834ca1d599\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817718 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c5ac722-cd40-44e9-921b-968dbe5e33f1-serving-cert\") pod \"service-ca-operator-777779d784-wgpqg\" (UID: \"4c5ac722-cd40-44e9-921b-968dbe5e33f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817746 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-plugins-dir\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817765 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92576279-7206-4601-9734-85243e90b5ed-config-volume\") pod \"dns-default-hc78d\" (UID: \"92576279-7206-4601-9734-85243e90b5ed\") " pod="openshift-dns/dns-default-hc78d" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817784 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wwc5\" (UniqueName: \"kubernetes.io/projected/4c5ac722-cd40-44e9-921b-968dbe5e33f1-kube-api-access-2wwc5\") pod \"service-ca-operator-777779d784-wgpqg\" (UID: \"4c5ac722-cd40-44e9-921b-968dbe5e33f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817801 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3cc1272d-e287-4e28-92f7-27dc0aa0d927-etcd-client\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817829 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghzsx\" (UniqueName: \"kubernetes.io/projected/143daef8-3687-4018-8f9f-2a675eac0909-kube-api-access-ghzsx\") pod \"marketplace-operator-79b997595-mbfc8\" (UID: \"143daef8-3687-4018-8f9f-2a675eac0909\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817849 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xknkv\" (UniqueName: \"kubernetes.io/projected/324debd2-da21-490a-9e83-36fbcb470aab-kube-api-access-xknkv\") pod \"machine-config-operator-74547568cd-ntvnm\" (UID: \"324debd2-da21-490a-9e83-36fbcb470aab\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817874 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76dnx\" (UniqueName: \"kubernetes.io/projected/9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450-kube-api-access-76dnx\") pod \"ingress-operator-5b745b69d9-lkssm\" (UID: \"9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817893 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjkt4\" (UniqueName: \"kubernetes.io/projected/041d1b1a-4861-4b7e-98ee-f5707404c173-kube-api-access-sjkt4\") pod \"control-plane-machine-set-operator-78cbb6b69f-55nsv\" (UID: \"041d1b1a-4861-4b7e-98ee-f5707404c173\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-55nsv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817916 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3cc1272d-e287-4e28-92f7-27dc0aa0d927-etcd-ca\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817937 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/041d1b1a-4861-4b7e-98ee-f5707404c173-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-55nsv\" (UID: \"041d1b1a-4861-4b7e-98ee-f5707404c173\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-55nsv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817955 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450-metrics-tls\") pod \"ingress-operator-5b745b69d9-lkssm\" (UID: \"9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817971 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/464c6029-4dac-48ee-9f4f-70659bb6445b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zmnzd\" (UID: \"464c6029-4dac-48ee-9f4f-70659bb6445b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.817998 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lkssm\" (UID: \"9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.818015 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-config-volume\") pod \"collect-profiles-29412645-6w78t\" (UID: \"d8e77566-f3a2-4a10-84b5-4e6078bc6f23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.818035 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/af7f7f9c-e126-413f-8d2d-499af6740741-profile-collector-cert\") pod \"catalog-operator-68c6474976-phkk4\" (UID: \"af7f7f9c-e126-413f-8d2d-499af6740741\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.818052 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6shw\" (UniqueName: \"kubernetes.io/projected/578c8a44-77e3-4c21-b841-3fc93a82b0f1-kube-api-access-g6shw\") pod \"package-server-manager-789f6589d5-2vh6v\" (UID: \"578c8a44-77e3-4c21-b841-3fc93a82b0f1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.818071 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9dc04fc6-6a79-4317-8b14-be963345a26c-proxy-tls\") pod \"machine-config-controller-84d6567774-cg4cm\" (UID: \"9dc04fc6-6a79-4317-8b14-be963345a26c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.818088 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cc1272d-e287-4e28-92f7-27dc0aa0d927-config\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.818101 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-secret-volume\") pod \"collect-profiles-29412645-6w78t\" (UID: \"d8e77566-f3a2-4a10-84b5-4e6078bc6f23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.818119 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9ztb\" (UniqueName: \"kubernetes.io/projected/75e92964-1833-4889-b175-16f4bba0ed1b-kube-api-access-q9ztb\") pod \"migrator-59844c95c7-prp8k\" (UID: \"75e92964-1833-4889-b175-16f4bba0ed1b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-prp8k" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.818134 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cc1272d-e287-4e28-92f7-27dc0aa0d927-serving-cert\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.818157 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cd573c9a-19ea-4975-9645-9d5b9e258d58-tmpfs\") pod \"packageserver-d55dfcdfc-9x4fj\" (UID: \"cd573c9a-19ea-4975-9645-9d5b9e258d58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.818534 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cd573c9a-19ea-4975-9645-9d5b9e258d58-tmpfs\") pod \"packageserver-d55dfcdfc-9x4fj\" (UID: \"cd573c9a-19ea-4975-9645-9d5b9e258d58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.818837 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-registration-dir\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.822378 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/92576279-7206-4601-9734-85243e90b5ed-metrics-tls\") pod \"dns-default-hc78d\" (UID: \"92576279-7206-4601-9734-85243e90b5ed\") " pod="openshift-dns/dns-default-hc78d" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.822892 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/324debd2-da21-490a-9e83-36fbcb470aab-images\") pod \"machine-config-operator-74547568cd-ntvnm\" (UID: \"324debd2-da21-490a-9e83-36fbcb470aab\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.824871 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c5ac722-cd40-44e9-921b-968dbe5e33f1-config\") pod \"service-ca-operator-777779d784-wgpqg\" (UID: \"4c5ac722-cd40-44e9-921b-968dbe5e33f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.826440 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/88293fb3-626a-4944-9d1f-2c569832661d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pkrnf\" (UID: \"88293fb3-626a-4944-9d1f-2c569832661d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.827265 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-csi-data-dir\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.831165 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3cc1272d-e287-4e28-92f7-27dc0aa0d927-etcd-ca\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.832039 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/578c8a44-77e3-4c21-b841-3fc93a82b0f1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2vh6v\" (UID: \"578c8a44-77e3-4c21-b841-3fc93a82b0f1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.832656 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-bound-sa-token\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.835634 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-config-volume\") pod \"collect-profiles-29412645-6w78t\" (UID: \"d8e77566-f3a2-4a10-84b5-4e6078bc6f23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.836304 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfvbk\" (UniqueName: \"kubernetes.io/projected/f20ff0d2-d132-4a30-83b5-e59f6e7260bf-kube-api-access-wfvbk\") pod \"cluster-samples-operator-665b6dd947-sc5lh\" (UID: \"f20ff0d2-d132-4a30-83b5-e59f6e7260bf\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sc5lh" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.837005 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9dc04fc6-6a79-4317-8b14-be963345a26c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-cg4cm\" (UID: \"9dc04fc6-6a79-4317-8b14-be963345a26c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.839387 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cc1272d-e287-4e28-92f7-27dc0aa0d927-config\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.839912 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-plugins-dir\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.847309 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40507d60-aae3-43da-85b0-09c02327ece8-cert\") pod \"ingress-canary-vtndh\" (UID: \"40507d60-aae3-43da-85b0-09c02327ece8\") " pod="openshift-ingress-canary/ingress-canary-vtndh" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.847819 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/682f390b-3453-40c5-aa8c-5a834ca1d599-config\") pod \"kube-controller-manager-operator-78b949d7b-5dz9v\" (UID: \"682f390b-3453-40c5-aa8c-5a834ca1d599\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.848088 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92576279-7206-4601-9734-85243e90b5ed-config-volume\") pod \"dns-default-hc78d\" (UID: \"92576279-7206-4601-9734-85243e90b5ed\") " pod="openshift-dns/dns-default-hc78d" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.848480 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/88293fb3-626a-4944-9d1f-2c569832661d-srv-cert\") pod \"olm-operator-6b444d44fb-pkrnf\" (UID: \"88293fb3-626a-4944-9d1f-2c569832661d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.848565 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-socket-dir\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.849626 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3cc1272d-e287-4e28-92f7-27dc0aa0d927-etcd-client\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.851196 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/324debd2-da21-490a-9e83-36fbcb470aab-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ntvnm\" (UID: \"324debd2-da21-490a-9e83-36fbcb470aab\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.852004 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa99705d-9e2f-4c32-85cb-04f0cde27411-service-ca-bundle\") pod \"router-default-5444994796-7whtv\" (UID: \"aa99705d-9e2f-4c32-85cb-04f0cde27411\") " pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:55:59 crc kubenswrapper[4646]: E1203 10:55:59.852463 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:00.352450104 +0000 UTC m=+136.815506239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.853161 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3cc1272d-e287-4e28-92f7-27dc0aa0d927-etcd-service-ca\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.853187 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aa99705d-9e2f-4c32-85cb-04f0cde27411-metrics-certs\") pod \"router-default-5444994796-7whtv\" (UID: \"aa99705d-9e2f-4c32-85cb-04f0cde27411\") " pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.853217 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-mountpoint-dir\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.853816 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cd573c9a-19ea-4975-9645-9d5b9e258d58-apiservice-cert\") pod \"packageserver-d55dfcdfc-9x4fj\" (UID: \"cd573c9a-19ea-4975-9645-9d5b9e258d58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.854353 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/45f218e8-2c99-412f-b85e-956417b5d74a-signing-cabundle\") pod \"service-ca-9c57cc56f-6rqmm\" (UID: \"45f218e8-2c99-412f-b85e-956417b5d74a\") " pod="openshift-service-ca/service-ca-9c57cc56f-6rqmm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.855430 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450-trusted-ca\") pod \"ingress-operator-5b745b69d9-lkssm\" (UID: \"9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.859191 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/464c6029-4dac-48ee-9f4f-70659bb6445b-config\") pod \"kube-apiserver-operator-766d6c64bb-zmnzd\" (UID: \"464c6029-4dac-48ee-9f4f-70659bb6445b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.862561 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/682f390b-3453-40c5-aa8c-5a834ca1d599-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5dz9v\" (UID: \"682f390b-3453-40c5-aa8c-5a834ca1d599\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.869453 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cd573c9a-19ea-4975-9645-9d5b9e258d58-webhook-cert\") pod \"packageserver-d55dfcdfc-9x4fj\" (UID: \"cd573c9a-19ea-4975-9645-9d5b9e258d58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.873041 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj"] Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.875614 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/324debd2-da21-490a-9e83-36fbcb470aab-proxy-tls\") pod \"machine-config-operator-74547568cd-ntvnm\" (UID: \"324debd2-da21-490a-9e83-36fbcb470aab\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.875707 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450-metrics-tls\") pod \"ingress-operator-5b745b69d9-lkssm\" (UID: \"9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.876155 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw9p8\" (UniqueName: \"kubernetes.io/projected/500b7283-f949-4ff4-a868-024bc3a8a6e5-kube-api-access-qw9p8\") pod \"cluster-image-registry-operator-dc59b4c8b-wndxj\" (UID: \"500b7283-f949-4ff4-a868-024bc3a8a6e5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" Dec 03 10:55:59 crc kubenswrapper[4646]: W1203 10:55:59.882904 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80cf56c0_7055_4e98_9fad_f728dd8b992a.slice/crio-5282cf75c90fd44cc8ff0a10928f1f6686299264f2aca23aec56283b4f2ac1b1 WatchSource:0}: Error finding container 5282cf75c90fd44cc8ff0a10928f1f6686299264f2aca23aec56283b4f2ac1b1: Status 404 returned error can't find the container with id 5282cf75c90fd44cc8ff0a10928f1f6686299264f2aca23aec56283b4f2ac1b1 Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.877595 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/464c6029-4dac-48ee-9f4f-70659bb6445b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zmnzd\" (UID: \"464c6029-4dac-48ee-9f4f-70659bb6445b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.883552 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9dc04fc6-6a79-4317-8b14-be963345a26c-proxy-tls\") pod \"machine-config-controller-84d6567774-cg4cm\" (UID: \"9dc04fc6-6a79-4317-8b14-be963345a26c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.883550 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c5ac722-cd40-44e9-921b-968dbe5e33f1-serving-cert\") pod \"service-ca-operator-777779d784-wgpqg\" (UID: \"4c5ac722-cd40-44e9-921b-968dbe5e33f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.885083 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/041d1b1a-4861-4b7e-98ee-f5707404c173-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-55nsv\" (UID: \"041d1b1a-4861-4b7e-98ee-f5707404c173\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-55nsv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.887306 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cc1272d-e287-4e28-92f7-27dc0aa0d927-serving-cert\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.888005 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-secret-volume\") pod \"collect-profiles-29412645-6w78t\" (UID: \"d8e77566-f3a2-4a10-84b5-4e6078bc6f23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.888284 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/af7f7f9c-e126-413f-8d2d-499af6740741-profile-collector-cert\") pod \"catalog-operator-68c6474976-phkk4\" (UID: \"af7f7f9c-e126-413f-8d2d-499af6740741\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.888316 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f9912d55-27ed-44cf-a597-9591fa35476c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hqnkq\" (UID: \"f9912d55-27ed-44cf-a597-9591fa35476c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hqnkq" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.888771 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/af7f7f9c-e126-413f-8d2d-499af6740741-srv-cert\") pod \"catalog-operator-68c6474976-phkk4\" (UID: \"af7f7f9c-e126-413f-8d2d-499af6740741\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.888871 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5wscd\" (UID: \"697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.890128 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/aa99705d-9e2f-4c32-85cb-04f0cde27411-default-certificate\") pod \"router-default-5444994796-7whtv\" (UID: \"aa99705d-9e2f-4c32-85cb-04f0cde27411\") " pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.890185 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/aa99705d-9e2f-4c32-85cb-04f0cde27411-stats-auth\") pod \"router-default-5444994796-7whtv\" (UID: \"aa99705d-9e2f-4c32-85cb-04f0cde27411\") " pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.892101 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4b507901-6faf-4fcf-a512-e876b56c54e1-certs\") pod \"machine-config-server-mvqf6\" (UID: \"4b507901-6faf-4fcf-a512-e876b56c54e1\") " pod="openshift-machine-config-operator/machine-config-server-mvqf6" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.893406 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4b507901-6faf-4fcf-a512-e876b56c54e1-node-bootstrap-token\") pod \"machine-config-server-mvqf6\" (UID: \"4b507901-6faf-4fcf-a512-e876b56c54e1\") " pod="openshift-machine-config-operator/machine-config-server-mvqf6" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.894643 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jcw4\" (UniqueName: \"kubernetes.io/projected/54c0778a-7e52-4da6-ab86-6a3492eb278b-kube-api-access-6jcw4\") pod \"downloads-7954f5f757-l2kck\" (UID: \"54c0778a-7e52-4da6-ab86-6a3492eb278b\") " pod="openshift-console/downloads-7954f5f757-l2kck" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.909861 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4qzw\" (UniqueName: \"kubernetes.io/projected/e6da1553-a91a-4055-af3d-bd30b20e6c0f-kube-api-access-x4qzw\") pod \"console-operator-58897d9998-52ndk\" (UID: \"e6da1553-a91a-4055-af3d-bd30b20e6c0f\") " pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.918873 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:55:59 crc kubenswrapper[4646]: E1203 10:55:59.919399 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:00.419382195 +0000 UTC m=+136.882438330 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.925763 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.926941 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xz4r\" (UniqueName: \"kubernetes.io/projected/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-kube-api-access-2xz4r\") pod \"console-f9d7485db-rvbtr\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.938588 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/500b7283-f949-4ff4-a868-024bc3a8a6e5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-wndxj\" (UID: \"500b7283-f949-4ff4-a868-024bc3a8a6e5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.944090 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q7bzw"] Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.948658 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sc5lh" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.962060 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq8vl\" (UniqueName: \"kubernetes.io/projected/c656cab0-e64d-4a92-96d5-3239ec762208-kube-api-access-hq8vl\") pod \"openshift-controller-manager-operator-756b6f6bc6-ztz2p\" (UID: \"c656cab0-e64d-4a92-96d5-3239ec762208\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p" Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.978824 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-lcfx6"] Dec 03 10:55:59 crc kubenswrapper[4646]: I1203 10:55:59.982646 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm9r5\" (UniqueName: \"kubernetes.io/projected/341c245b-3af9-42fd-9592-00737bcc7ba7-kube-api-access-rm9r5\") pod \"kube-storage-version-migrator-operator-b67b599dd-cb8tj\" (UID: \"341c245b-3af9-42fd-9592-00737bcc7ba7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.005939 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwtbr\" (UniqueName: \"kubernetes.io/projected/3cc1272d-e287-4e28-92f7-27dc0aa0d927-kube-api-access-cwtbr\") pod \"etcd-operator-b45778765-4ftgs\" (UID: \"3cc1272d-e287-4e28-92f7-27dc0aa0d927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.013656 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.019788 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjwhf\" (UniqueName: \"kubernetes.io/projected/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-kube-api-access-tjwhf\") pod \"collect-profiles-29412645-6w78t\" (UID: \"d8e77566-f3a2-4a10-84b5-4e6078bc6f23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.020518 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:00 crc kubenswrapper[4646]: E1203 10:56:00.020899 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:00.520882802 +0000 UTC m=+136.983938937 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.038092 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-697vl"] Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.045223 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dswf8\" (UniqueName: \"kubernetes.io/projected/af7f7f9c-e126-413f-8d2d-499af6740741-kube-api-access-dswf8\") pod \"catalog-operator-68c6474976-phkk4\" (UID: \"af7f7f9c-e126-413f-8d2d-499af6740741\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.058027 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2trtv"] Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.062085 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.064859 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bhzv\" (UniqueName: \"kubernetes.io/projected/92576279-7206-4601-9734-85243e90b5ed-kube-api-access-8bhzv\") pod \"dns-default-hc78d\" (UID: \"92576279-7206-4601-9734-85243e90b5ed\") " pod="openshift-dns/dns-default-hc78d" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.079233 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4f9n\" (UniqueName: \"kubernetes.io/projected/dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6-kube-api-access-m4f9n\") pod \"csi-hostpathplugin-czsqx\" (UID: \"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6\") " pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.083779 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.083788 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-l2kck" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.094318 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-4jgxv"] Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.098281 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5wscd\" (UID: \"697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.099123 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.121425 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh"] Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.121750 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:00 crc kubenswrapper[4646]: E1203 10:56:00.121894 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:00.621873555 +0000 UTC m=+137.084929690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.121988 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:00 crc kubenswrapper[4646]: E1203 10:56:00.122308 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:00.622297727 +0000 UTC m=+137.085353862 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.124073 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xknkv\" (UniqueName: \"kubernetes.io/projected/324debd2-da21-490a-9e83-36fbcb470aab-kube-api-access-xknkv\") pod \"machine-config-operator-74547568cd-ntvnm\" (UID: \"324debd2-da21-490a-9e83-36fbcb470aab\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.137255 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76dnx\" (UniqueName: \"kubernetes.io/projected/9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450-kube-api-access-76dnx\") pod \"ingress-operator-5b745b69d9-lkssm\" (UID: \"9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.145087 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.157879 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjkt4\" (UniqueName: \"kubernetes.io/projected/041d1b1a-4861-4b7e-98ee-f5707404c173-kube-api-access-sjkt4\") pod \"control-plane-machine-set-operator-78cbb6b69f-55nsv\" (UID: \"041d1b1a-4861-4b7e-98ee-f5707404c173\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-55nsv" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.159683 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.172087 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.197638 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/682f390b-3453-40c5-aa8c-5a834ca1d599-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5dz9v\" (UID: \"682f390b-3453-40c5-aa8c-5a834ca1d599\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.216976 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.217642 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbjz4\" (UniqueName: \"kubernetes.io/projected/9dc04fc6-6a79-4317-8b14-be963345a26c-kube-api-access-mbjz4\") pod \"machine-config-controller-84d6567774-cg4cm\" (UID: \"9dc04fc6-6a79-4317-8b14-be963345a26c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.222851 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:00 crc kubenswrapper[4646]: E1203 10:56:00.223177 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:00.723165297 +0000 UTC m=+137.186221432 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.236278 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.238472 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xsm7\" (UniqueName: \"kubernetes.io/projected/88293fb3-626a-4944-9d1f-2c569832661d-kube-api-access-4xsm7\") pod \"olm-operator-6b444d44fb-pkrnf\" (UID: \"88293fb3-626a-4944-9d1f-2c569832661d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.249287 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-55nsv" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.263166 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hc78d" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.278700 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-czsqx" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.281803 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lkssm\" (UID: \"9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.303743 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9ztb\" (UniqueName: \"kubernetes.io/projected/75e92964-1833-4889-b175-16f4bba0ed1b-kube-api-access-q9ztb\") pod \"migrator-59844c95c7-prp8k\" (UID: \"75e92964-1833-4889-b175-16f4bba0ed1b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-prp8k" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.324392 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:00 crc kubenswrapper[4646]: E1203 10:56:00.324788 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:00.824769457 +0000 UTC m=+137.287825602 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.330279 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wwc5\" (UniqueName: \"kubernetes.io/projected/4c5ac722-cd40-44e9-921b-968dbe5e33f1-kube-api-access-2wwc5\") pod \"service-ca-operator-777779d784-wgpqg\" (UID: \"4c5ac722-cd40-44e9-921b-968dbe5e33f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.344127 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghzsx\" (UniqueName: \"kubernetes.io/projected/143daef8-3687-4018-8f9f-2a675eac0909-kube-api-access-ghzsx\") pod \"marketplace-operator-79b997595-mbfc8\" (UID: \"143daef8-3687-4018-8f9f-2a675eac0909\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.364003 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmdqv\" (UniqueName: \"kubernetes.io/projected/aa99705d-9e2f-4c32-85cb-04f0cde27411-kube-api-access-kmdqv\") pod \"router-default-5444994796-7whtv\" (UID: \"aa99705d-9e2f-4c32-85cb-04f0cde27411\") " pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.383299 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwsqc\" (UniqueName: \"kubernetes.io/projected/4b507901-6faf-4fcf-a512-e876b56c54e1-kube-api-access-kwsqc\") pod \"machine-config-server-mvqf6\" (UID: \"4b507901-6faf-4fcf-a512-e876b56c54e1\") " pod="openshift-machine-config-operator/machine-config-server-mvqf6" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.400014 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5wscd\" (UID: \"697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.400730 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/143daef8-3687-4018-8f9f-2a675eac0909-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mbfc8\" (UID: \"143daef8-3687-4018-8f9f-2a675eac0909\") " pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.403188 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6shw\" (UniqueName: \"kubernetes.io/projected/578c8a44-77e3-4c21-b841-3fc93a82b0f1-kube-api-access-g6shw\") pod \"package-server-manager-789f6589d5-2vh6v\" (UID: \"578c8a44-77e3-4c21-b841-3fc93a82b0f1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.407521 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9p4v\" (UniqueName: \"kubernetes.io/projected/cd573c9a-19ea-4975-9645-9d5b9e258d58-kube-api-access-f9p4v\") pod \"packageserver-d55dfcdfc-9x4fj\" (UID: \"cd573c9a-19ea-4975-9645-9d5b9e258d58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.409852 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rghjj\" (UniqueName: \"kubernetes.io/projected/45f218e8-2c99-412f-b85e-956417b5d74a-kube-api-access-rghjj\") pod \"service-ca-9c57cc56f-6rqmm\" (UID: \"45f218e8-2c99-412f-b85e-956417b5d74a\") " pod="openshift-service-ca/service-ca-9c57cc56f-6rqmm" Dec 03 10:56:00 crc kubenswrapper[4646]: W1203 10:56:00.421100 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cc84826_c569_408e_94e9_8cf9bc49eae9.slice/crio-bde3dd835ddc7ad20c2d8fab4612fec11286cb3a641b25973bf593dc6005dc01 WatchSource:0}: Error finding container bde3dd835ddc7ad20c2d8fab4612fec11286cb3a641b25973bf593dc6005dc01: Status 404 returned error can't find the container with id bde3dd835ddc7ad20c2d8fab4612fec11286cb3a641b25973bf593dc6005dc01 Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.421399 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.425697 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:00 crc kubenswrapper[4646]: E1203 10:56:00.426047 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:00.926033288 +0000 UTC m=+137.389089413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.428670 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.429365 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz558\" (UniqueName: \"kubernetes.io/projected/f9912d55-27ed-44cf-a597-9591fa35476c-kube-api-access-tz558\") pod \"multus-admission-controller-857f4d67dd-hqnkq\" (UID: \"f9912d55-27ed-44cf-a597-9591fa35476c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hqnkq" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.435607 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.451707 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77pxz\" (UniqueName: \"kubernetes.io/projected/40507d60-aae3-43da-85b0-09c02327ece8-kube-api-access-77pxz\") pod \"ingress-canary-vtndh\" (UID: \"40507d60-aae3-43da-85b0-09c02327ece8\") " pod="openshift-ingress-canary/ingress-canary-vtndh" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.452011 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.462775 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/464c6029-4dac-48ee-9f4f-70659bb6445b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zmnzd\" (UID: \"464c6029-4dac-48ee-9f4f-70659bb6445b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.466690 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.479150 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-hqnkq" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.487358 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.493126 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.501758 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-prp8k" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.509413 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6rqmm" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.516068 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.523590 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-vtndh" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.526937 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:00 crc kubenswrapper[4646]: E1203 10:56:00.527246 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:01.027225308 +0000 UTC m=+137.490281443 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.529573 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.543513 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.557677 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-mvqf6" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.627984 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:00 crc kubenswrapper[4646]: E1203 10:56:00.628462 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:01.128442837 +0000 UTC m=+137.591498972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.656296 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" event={"ID":"1cc84826-c569-408e-94e9-8cf9bc49eae9","Type":"ContainerStarted","Data":"bde3dd835ddc7ad20c2d8fab4612fec11286cb3a641b25973bf593dc6005dc01"} Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.667446 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" event={"ID":"f653a1d6-4938-418c-9b40-ad23d437a156","Type":"ContainerStarted","Data":"ef4ac49e90386ee987c1a010e013cc96b29a682640e418f3d38918c70a09c793"} Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.674721 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" event={"ID":"aa151fb7-08d1-48b4-b477-e3310ff2274d","Type":"ContainerStarted","Data":"acd0b04c51e5579ffaa02e607fb215a9045a305d50acaece66436a390899feb5"} Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.675819 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" event={"ID":"e51eada8-802d-4064-8705-2647ffa44189","Type":"ContainerStarted","Data":"dff929edc30561bd1de5123e47771db8e9af2fbabaea5811a77db03ccfbe01de"} Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.676666 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-4jgxv" event={"ID":"61e9e161-fb19-4da8-9e38-4ca880134e94","Type":"ContainerStarted","Data":"624fced0aacf04addfc903b1880e9d4db03bcc9a9010a464f5f6b6d7035bd960"} Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.694758 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" event={"ID":"53824d3c-1dd1-4f25-b699-de9230490c84","Type":"ContainerStarted","Data":"63fb9684599a8e3573d0a4f94088c0c5cab75b002fee517ca06ce712d052f5cc"} Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.704173 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" event={"ID":"80cf56c0-7055-4e98-9fad-f728dd8b992a","Type":"ContainerStarted","Data":"5282cf75c90fd44cc8ff0a10928f1f6686299264f2aca23aec56283b4f2ac1b1"} Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.712960 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh" event={"ID":"c8936a9d-f33c-4de3-a759-7df7cf3e48b2","Type":"ContainerStarted","Data":"73bedff6766f871dc9aee64fd8996443d1b004fe08a2099d12fb36aa82596b86"} Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.713749 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd" Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.724004 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nb6wv"] Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.729073 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:00 crc kubenswrapper[4646]: E1203 10:56:00.729960 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:01.229942025 +0000 UTC m=+137.692998160 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.734512 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" event={"ID":"c202a4cb-ebe8-419d-a42d-27a286eeb4a0","Type":"ContainerStarted","Data":"c6970f1389208c55e73c9f8ac50a837df33a8a2b74ad6b375bedf38605467bf5"} Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.749650 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" event={"ID":"9a0dc011-d040-4ec5-babf-d95b03ec699e","Type":"ContainerStarted","Data":"0d06913b4fbb0f2ecaf7beb9f219eed6fc9d06990dc9c0e0785e48e9241bd905"} Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.808010 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-rvbtr"] Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.830115 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:00 crc kubenswrapper[4646]: E1203 10:56:00.830480 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:01.330448835 +0000 UTC m=+137.793505040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.835502 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sc5lh"] Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.875637 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4"] Dec 03 10:56:00 crc kubenswrapper[4646]: I1203 10:56:00.932478 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:00 crc kubenswrapper[4646]: E1203 10:56:00.932772 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:01.432761066 +0000 UTC m=+137.895817201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:01 crc kubenswrapper[4646]: W1203 10:56:01.031046 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd109aa7c_89de_44b3_852c_b3c49b2bb2a0.slice/crio-713f4c66c4c1ce1b2088c92f93b18de18ee2a87d78da93afc25d157349909056 WatchSource:0}: Error finding container 713f4c66c4c1ce1b2088c92f93b18de18ee2a87d78da93afc25d157349909056: Status 404 returned error can't find the container with id 713f4c66c4c1ce1b2088c92f93b18de18ee2a87d78da93afc25d157349909056 Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.034687 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:01 crc kubenswrapper[4646]: E1203 10:56:01.034849 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:01.534821059 +0000 UTC m=+137.997877184 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.034927 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:01 crc kubenswrapper[4646]: E1203 10:56:01.035231 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:01.535220881 +0000 UTC m=+137.998277016 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.046622 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p"] Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.077611 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-l2kck"] Dec 03 10:56:01 crc kubenswrapper[4646]: W1203 10:56:01.090007 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa99705d_9e2f_4c32_85cb_04f0cde27411.slice/crio-aa60e4ab87ea421816eff1f3de45d7fb808e714a3f3954c34cecfb8bcd3aa9f7 WatchSource:0}: Error finding container aa60e4ab87ea421816eff1f3de45d7fb808e714a3f3954c34cecfb8bcd3aa9f7: Status 404 returned error can't find the container with id aa60e4ab87ea421816eff1f3de45d7fb808e714a3f3954c34cecfb8bcd3aa9f7 Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.137828 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:01 crc kubenswrapper[4646]: E1203 10:56:01.138470 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:01.638453178 +0000 UTC m=+138.101509313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.240297 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:01 crc kubenswrapper[4646]: E1203 10:56:01.240901 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:01.740886113 +0000 UTC m=+138.203942248 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.342125 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:01 crc kubenswrapper[4646]: E1203 10:56:01.342601 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:01.842586486 +0000 UTC m=+138.305642621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.449273 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:01 crc kubenswrapper[4646]: E1203 10:56:01.449631 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:01.949619142 +0000 UTC m=+138.412675267 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.550807 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:01 crc kubenswrapper[4646]: E1203 10:56:01.551258 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:02.051213443 +0000 UTC m=+138.514269578 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.551632 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:01 crc kubenswrapper[4646]: E1203 10:56:01.552711 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:02.052682205 +0000 UTC m=+138.515738340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.653141 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:01 crc kubenswrapper[4646]: E1203 10:56:01.653360 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:02.153316348 +0000 UTC m=+138.616372483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.653483 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:01 crc kubenswrapper[4646]: E1203 10:56:01.654028 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:02.154016648 +0000 UTC m=+138.617072783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.671118 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm"] Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.679229 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-czsqx"] Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.755952 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:01 crc kubenswrapper[4646]: E1203 10:56:01.756312 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:02.256297688 +0000 UTC m=+138.719353823 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.794705 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7whtv" event={"ID":"aa99705d-9e2f-4c32-85cb-04f0cde27411","Type":"ContainerStarted","Data":"aa60e4ab87ea421816eff1f3de45d7fb808e714a3f3954c34cecfb8bcd3aa9f7"} Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.805286 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh" event={"ID":"c8936a9d-f33c-4de3-a759-7df7cf3e48b2","Type":"ContainerStarted","Data":"758d327de52f7cfd63a8edef579670e591b33704f786bd55b5d0da08b5466de0"} Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.806856 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" event={"ID":"80cf56c0-7055-4e98-9fad-f728dd8b992a","Type":"ContainerStarted","Data":"53310597ee480a18ec896d75cd1e1bb35d44fb4cd8f9ea335c8b84f38cd01943"} Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.806930 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj"] Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.807197 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.842646 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rvbtr" event={"ID":"d109aa7c-89de-44b3-852c-b3c49b2bb2a0","Type":"ContainerStarted","Data":"713f4c66c4c1ce1b2088c92f93b18de18ee2a87d78da93afc25d157349909056"} Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.861572 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:01 crc kubenswrapper[4646]: E1203 10:56:01.861855 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:02.361844951 +0000 UTC m=+138.824901086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.875664 4646 generic.go:334] "Generic (PLEG): container finished" podID="c202a4cb-ebe8-419d-a42d-27a286eeb4a0" containerID="7586b9d3b1a202dffec332665692e2bbf68a9ac0c47bbff4a3b8178469d14391" exitCode=0 Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.919203 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj"] Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.919231 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" event={"ID":"c202a4cb-ebe8-419d-a42d-27a286eeb4a0","Type":"ContainerDied","Data":"7586b9d3b1a202dffec332665692e2bbf68a9ac0c47bbff4a3b8178469d14391"} Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.919255 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-55nsv"] Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.919266 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd"] Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.926478 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hc78d"] Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.933227 4646 generic.go:334] "Generic (PLEG): container finished" podID="9a0dc011-d040-4ec5-babf-d95b03ec699e" containerID="8aa572e153845c0498de4d1df679a3e35cc3d1e219462dcabf73a67c3fd7c119" exitCode=0 Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.933498 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" event={"ID":"9a0dc011-d040-4ec5-babf-d95b03ec699e","Type":"ContainerDied","Data":"8aa572e153845c0498de4d1df679a3e35cc3d1e219462dcabf73a67c3fd7c119"} Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.939682 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t"] Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.963616 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:01 crc kubenswrapper[4646]: I1203 10:56:01.985043 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" event={"ID":"aa151fb7-08d1-48b4-b477-e3310ff2274d","Type":"ContainerStarted","Data":"12eef9e3f6926afd7cdc537b9bbadae60c3a86b55929f7416174acc14785f60c"} Dec 03 10:56:02 crc kubenswrapper[4646]: E1203 10:56:02.012613 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:02.512577765 +0000 UTC m=+138.975633900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.023093 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g2szh" podStartSLOduration=120.023071534 podStartE2EDuration="2m0.023071534s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:01.984711439 +0000 UTC m=+138.447767574" watchObservedRunningTime="2025-12-03 10:56:02.023071534 +0000 UTC m=+138.486127669" Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.039152 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-l2kck" event={"ID":"54c0778a-7e52-4da6-ab86-6a3492eb278b","Type":"ContainerStarted","Data":"7d82cd843ea78923aa1d23a759267f04e1fc7bd92f680b9b72d715073b963261"} Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.085569 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" podStartSLOduration=119.085551318 podStartE2EDuration="1m59.085551318s" podCreationTimestamp="2025-12-03 10:54:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:02.083813489 +0000 UTC m=+138.546869624" watchObservedRunningTime="2025-12-03 10:56:02.085551318 +0000 UTC m=+138.548607453" Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.098468 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:02 crc kubenswrapper[4646]: E1203 10:56:02.099491 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:02.599480056 +0000 UTC m=+139.062536191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.164599 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-mvqf6" event={"ID":"4b507901-6faf-4fcf-a512-e876b56c54e1","Type":"ContainerStarted","Data":"d625aff7faa5a10962c12a345b479e49055d99d6ea6bc448bfaf545b0d29a9a1"} Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.179928 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-2trtv" podStartSLOduration=120.179906882 podStartE2EDuration="2m0.179906882s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:02.135381861 +0000 UTC m=+138.598437996" watchObservedRunningTime="2025-12-03 10:56:02.179906882 +0000 UTC m=+138.642963017" Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.200010 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:02 crc kubenswrapper[4646]: E1203 10:56:02.200431 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:02.700411168 +0000 UTC m=+139.163467303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.206733 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:02 crc kubenswrapper[4646]: E1203 10:56:02.207097 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:02.707083668 +0000 UTC m=+139.170139803 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.228938 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" event={"ID":"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924","Type":"ContainerStarted","Data":"622bc5cf6198552c568236318014324bcba07c598f9e6704958cb17521ce03c9"} Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.260156 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" event={"ID":"af7f7f9c-e126-413f-8d2d-499af6740741","Type":"ContainerStarted","Data":"e2bf0061f72b0a0ab5e8008f7bf4b7058c59f00182e9686595e637df45a8b063"} Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.276434 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sc5lh" event={"ID":"f20ff0d2-d132-4a30-83b5-e59f6e7260bf","Type":"ContainerStarted","Data":"abdd995838c00cbcccff983e9647a46288dd527859a4c26f200bf3ba3642165f"} Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.281623 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" event={"ID":"1cc84826-c569-408e-94e9-8cf9bc49eae9","Type":"ContainerStarted","Data":"fc7637bb6c13ba2fd6a2f7e9756c74a8d7b0a70623564002c2ceb4bef3dd7948"} Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.301972 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" event={"ID":"e51eada8-802d-4064-8705-2647ffa44189","Type":"ContainerStarted","Data":"f5b280f3f57a36058f5aa0b2f298d810688b521848dda8b6ce675fe80b058e23"} Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.311757 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:02 crc kubenswrapper[4646]: E1203 10:56:02.312143 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:02.812127017 +0000 UTC m=+139.275183152 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.316115 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.325383 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p" event={"ID":"c656cab0-e64d-4a92-96d5-3239ec762208","Type":"ContainerStarted","Data":"5ee06e7ca802c068537097fa1482b87242ad92198557d5947e77c78fc71bb08b"} Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.335328 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.365261 4646 generic.go:334] "Generic (PLEG): container finished" podID="53824d3c-1dd1-4f25-b699-de9230490c84" containerID="bb57d60d424c7a1e349d1499d808c5136cffe98e0ac6a56da18800eb06e95cc2" exitCode=0 Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.365485 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" event={"ID":"53824d3c-1dd1-4f25-b699-de9230490c84","Type":"ContainerDied","Data":"bb57d60d424c7a1e349d1499d808c5136cffe98e0ac6a56da18800eb06e95cc2"} Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.413827 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:02 crc kubenswrapper[4646]: E1203 10:56:02.415394 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:02.915381895 +0000 UTC m=+139.378438030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.431922 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm"] Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.467590 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" podStartSLOduration=120.467575625 podStartE2EDuration="2m0.467575625s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:02.467545084 +0000 UTC m=+138.930601219" watchObservedRunningTime="2025-12-03 10:56:02.467575625 +0000 UTC m=+138.930631760" Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.514722 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:02 crc kubenswrapper[4646]: E1203 10:56:02.515641 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:03.015623797 +0000 UTC m=+139.478679932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.619480 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:02 crc kubenswrapper[4646]: E1203 10:56:02.620093 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:03.120077069 +0000 UTC m=+139.583133204 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.722719 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:02 crc kubenswrapper[4646]: E1203 10:56:02.722884 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:03.222868834 +0000 UTC m=+139.685924959 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.722919 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:02 crc kubenswrapper[4646]: E1203 10:56:02.723218 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:03.223210663 +0000 UTC m=+139.686266798 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.771921 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.774947 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-52ndk"] Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.826765 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:02 crc kubenswrapper[4646]: E1203 10:56:02.827079 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:03.327065479 +0000 UTC m=+139.790121614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.910181 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf"] Dec 03 10:56:02 crc kubenswrapper[4646]: I1203 10:56:02.954458 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:02 crc kubenswrapper[4646]: E1203 10:56:02.954916 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:03.454902698 +0000 UTC m=+139.917958833 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.046156 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v"] Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.055283 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:03 crc kubenswrapper[4646]: E1203 10:56:03.055807 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:03.555791099 +0000 UTC m=+140.018847234 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.081923 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mbfc8"] Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.160132 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:03 crc kubenswrapper[4646]: E1203 10:56:03.161348 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:03.661318432 +0000 UTC m=+140.124374567 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.251250 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm"] Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.263042 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:03 crc kubenswrapper[4646]: E1203 10:56:03.263644 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:03.763620813 +0000 UTC m=+140.226676948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.283602 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-vtndh"] Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.364601 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:03 crc kubenswrapper[4646]: E1203 10:56:03.364985 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:03.864972946 +0000 UTC m=+140.328029081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.411644 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" event={"ID":"143daef8-3687-4018-8f9f-2a675eac0909","Type":"ContainerStarted","Data":"f9b27d11775849302b28543ffe31fe8721e2ab8538efaa59f7c8c40be1ae7e33"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.413956 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v" event={"ID":"578c8a44-77e3-4c21-b841-3fc93a82b0f1","Type":"ContainerStarted","Data":"39f625de2d97be72df4289ea7c65fa3c8a0f9763d642de4ea5b03ec1b048620e"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.415314 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm" event={"ID":"9dc04fc6-6a79-4317-8b14-be963345a26c","Type":"ContainerStarted","Data":"11b36f9e05bfececc045a759e896d97dde96bd0f76b4c2e3ff4c3f58a543790c"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.445031 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" event={"ID":"d8e77566-f3a2-4a10-84b5-4e6078bc6f23","Type":"ContainerStarted","Data":"011d2186daa42776238b6189b1866584dcfdc7148abe0731e3b9c1021271b63d"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.445074 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" event={"ID":"d8e77566-f3a2-4a10-84b5-4e6078bc6f23","Type":"ContainerStarted","Data":"d895cd81faca6bd4c4ff8eccaf84bf9257310e4b9166319dc9451e27ee4da175"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.466210 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:03 crc kubenswrapper[4646]: E1203 10:56:03.466829 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:03.966806094 +0000 UTC m=+140.429862229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.466943 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:03 crc kubenswrapper[4646]: E1203 10:56:03.468061 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:03.968045259 +0000 UTC m=+140.431101394 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.484078 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7whtv" event={"ID":"aa99705d-9e2f-4c32-85cb-04f0cde27411","Type":"ContainerStarted","Data":"26188528a9ea5292f06ae18c745799aca6ae46aac1173070e2efc7126f81180f"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.486937 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj"] Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.489019 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd"] Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.508040 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-52ndk" event={"ID":"e6da1553-a91a-4055-af3d-bd30b20e6c0f","Type":"ContainerStarted","Data":"630b79adc0d380831ef177a53bf4fac867f1bd04f3f56ecfb6a3f46138ec7fb3"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.511831 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg"] Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.519400 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" podStartSLOduration=122.519383315 podStartE2EDuration="2m2.519383315s" podCreationTimestamp="2025-12-03 10:54:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:03.518742787 +0000 UTC m=+139.981798922" watchObservedRunningTime="2025-12-03 10:56:03.519383315 +0000 UTC m=+139.982439450" Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.548045 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" event={"ID":"1cc84826-c569-408e-94e9-8cf9bc49eae9","Type":"ContainerStarted","Data":"8eeeb532bb8cf5d0c5c19127b54e2f87a9533378d5e72da9d9a6b8b19db277da"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.563486 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-prp8k"] Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.563541 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v"] Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.577660 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:03 crc kubenswrapper[4646]: E1203 10:56:03.577757 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:04.077734751 +0000 UTC m=+140.540790886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.577953 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:03 crc kubenswrapper[4646]: E1203 10:56:03.578232 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:04.078217785 +0000 UTC m=+140.541273920 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.599908 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-7whtv" podStartSLOduration=121.599894652 podStartE2EDuration="2m1.599894652s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:03.598582955 +0000 UTC m=+140.061639090" watchObservedRunningTime="2025-12-03 10:56:03.599894652 +0000 UTC m=+140.062950787" Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.614707 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rvbtr" event={"ID":"d109aa7c-89de-44b3-852c-b3c49b2bb2a0","Type":"ContainerStarted","Data":"0a2d8ecb23e9e05e7a1254d63cb75893768cf77b9b1bbe953cefbd44d09a45c9"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.622221 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" event={"ID":"f653a1d6-4938-418c-9b40-ad23d437a156","Type":"ContainerStarted","Data":"ccc07c2276dc557e2ca6b481e543be9bd66a98ee6e78af8f7219e19585a28470"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.627356 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-mvqf6" event={"ID":"4b507901-6faf-4fcf-a512-e876b56c54e1","Type":"ContainerStarted","Data":"684fb9a3aebe7804affb54f1803d1f7ba7691cf9251a8ae02c9c2d99e8d0d051"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.628394 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" event={"ID":"88293fb3-626a-4944-9d1f-2c569832661d","Type":"ContainerStarted","Data":"670a3c21a8b9526a98ce88c9df0906425f725ea4c073bb19a6a7d60fe2a159af"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.629179 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" event={"ID":"af7f7f9c-e126-413f-8d2d-499af6740741","Type":"ContainerStarted","Data":"b442a5b68d2e8294279c572d07fd5e4f7bd207b706b89bacf76b658bd63539a4"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.629767 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.630379 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-czsqx" event={"ID":"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6","Type":"ContainerStarted","Data":"c3d32dc845e83323f622d5e3027fc6daba03dc09d73dc65aeea44eb18b6c9b6b"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.631173 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hc78d" event={"ID":"92576279-7206-4601-9734-85243e90b5ed","Type":"ContainerStarted","Data":"660b47dc223a225756a55c14230d7f3ce76e0bcaa2496f3397c7df63ef21e185"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.646466 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.647996 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" event={"ID":"500b7283-f949-4ff4-a868-024bc3a8a6e5","Type":"ContainerStarted","Data":"70b6850c2015a0bce68c448f9ec8d8e1bc16f08ebdbe9dfa61746bfe8b9fed2e"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.658304 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sc5lh" event={"ID":"f20ff0d2-d132-4a30-83b5-e59f6e7260bf","Type":"ContainerStarted","Data":"19b1df0ae8286dfc1c4b276d11e4d252890bbd9f2ae3300601891572dc75d1bc"} Dec 03 10:56:03 crc kubenswrapper[4646]: W1203 10:56:03.669579 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c5ac722_cd40_44e9_921b_968dbe5e33f1.slice/crio-e285b7f84fea1da1e91ab690ffeab6a25bdf942e776f7e27ae48d5658ec317a5 WatchSource:0}: Error finding container e285b7f84fea1da1e91ab690ffeab6a25bdf942e776f7e27ae48d5658ec317a5: Status 404 returned error can't find the container with id e285b7f84fea1da1e91ab690ffeab6a25bdf942e776f7e27ae48d5658ec317a5 Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.670100 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj" event={"ID":"341c245b-3af9-42fd-9592-00737bcc7ba7","Type":"ContainerStarted","Data":"e8e59591903c35f385448111e3d6e06135640d7d672259cdc2f91522e5b30271"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.676074 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-lcfx6" podStartSLOduration=121.676060387 podStartE2EDuration="2m1.676060387s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:03.646118452 +0000 UTC m=+140.109174587" watchObservedRunningTime="2025-12-03 10:56:03.676060387 +0000 UTC m=+140.139116522" Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.682689 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:03 crc kubenswrapper[4646]: E1203 10:56:03.684039 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:04.184016744 +0000 UTC m=+140.647072879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.693972 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-4jgxv" event={"ID":"61e9e161-fb19-4da8-9e38-4ca880134e94","Type":"ContainerStarted","Data":"032ea264258828f3c0d4bdc5dc03479572c422926126ae7079251a24601a871d"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.720659 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-mvqf6" podStartSLOduration=6.72064593 podStartE2EDuration="6.72064593s" podCreationTimestamp="2025-12-03 10:55:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:03.677498208 +0000 UTC m=+140.140554343" watchObservedRunningTime="2025-12-03 10:56:03.72064593 +0000 UTC m=+140.183702065" Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.741123 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-rvbtr" podStartSLOduration=121.741104074 podStartE2EDuration="2m1.741104074s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:03.719471756 +0000 UTC m=+140.182527891" watchObservedRunningTime="2025-12-03 10:56:03.741104074 +0000 UTC m=+140.204160209" Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.741611 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-4ftgs"] Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.756468 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd" event={"ID":"697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed","Type":"ContainerStarted","Data":"4309ed66881a8a51c71d376c7f42f8c923a7460f0fc3b207f7cb4e4bad23efa4"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.780382 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-l2kck" event={"ID":"54c0778a-7e52-4da6-ab86-6a3492eb278b","Type":"ContainerStarted","Data":"d732563cf358efd87adc6862c3b7a3f9e480b3336017c1f5caa27e3726d2e6f6"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.781448 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-l2kck" Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.782944 4646 patch_prober.go:28] interesting pod/downloads-7954f5f757-l2kck container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.783083 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-l2kck" podUID="54c0778a-7e52-4da6-ab86-6a3492eb278b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.784017 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:03 crc kubenswrapper[4646]: E1203 10:56:03.785318 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:04.285307516 +0000 UTC m=+140.748363651 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.785822 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-phkk4" podStartSLOduration=121.7857984 podStartE2EDuration="2m1.7857984s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:03.770668188 +0000 UTC m=+140.233724323" watchObservedRunningTime="2025-12-03 10:56:03.7857984 +0000 UTC m=+140.248854535" Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.786368 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6rqmm"] Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.805683 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-55nsv" event={"ID":"041d1b1a-4861-4b7e-98ee-f5707404c173","Type":"ContainerStarted","Data":"5cc93dd835b1f7a90a3d93ef01e43e22a044e8c8d7a521707790591956a1341e"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.812658 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p7xvl" podStartSLOduration=122.812641526 podStartE2EDuration="2m2.812641526s" podCreationTimestamp="2025-12-03 10:54:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:03.811908945 +0000 UTC m=+140.274965070" watchObservedRunningTime="2025-12-03 10:56:03.812641526 +0000 UTC m=+140.275697661" Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.836037 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" event={"ID":"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924","Type":"ContainerStarted","Data":"41f8ff013e88e5457e4812f84d955623880c35a81841c74e9cb7e20a190bbb68"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.836252 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.859472 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" event={"ID":"324debd2-da21-490a-9e83-36fbcb470aab","Type":"ContainerStarted","Data":"497cc5170b0912fab183b15a1c1f49ba60772a6d3a77841d96f849a34bf3488a"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.859611 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" event={"ID":"324debd2-da21-490a-9e83-36fbcb470aab","Type":"ContainerStarted","Data":"a54a897f89d54ee849a592dc48ad5bc743fcae63d49ddc4b2f1659d5033c2e70"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.877512 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p" event={"ID":"c656cab0-e64d-4a92-96d5-3239ec762208","Type":"ContainerStarted","Data":"98e2b99f8177cd6d9753aa0a075eb602c6546e9937fa5f38b4adde0483184965"} Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.886738 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:03 crc kubenswrapper[4646]: E1203 10:56:03.887685 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:04.387658708 +0000 UTC m=+140.850714843 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.932460 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-l2kck" podStartSLOduration=121.932438297 podStartE2EDuration="2m1.932438297s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:03.925348654 +0000 UTC m=+140.388404789" watchObservedRunningTime="2025-12-03 10:56:03.932438297 +0000 UTC m=+140.395494432" Dec 03 10:56:03 crc kubenswrapper[4646]: I1203 10:56:03.988220 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:04 crc kubenswrapper[4646]: E1203 10:56:04.006355 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:04.506320366 +0000 UTC m=+140.969376501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.024442 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hqnkq"] Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.042981 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd" podStartSLOduration=122.042958682 podStartE2EDuration="2m2.042958682s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:03.972977584 +0000 UTC m=+140.436033709" watchObservedRunningTime="2025-12-03 10:56:04.042958682 +0000 UTC m=+140.506014807" Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.076967 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" podStartSLOduration=122.076946842 podStartE2EDuration="2m2.076946842s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:04.067148242 +0000 UTC m=+140.530204377" watchObservedRunningTime="2025-12-03 10:56:04.076946842 +0000 UTC m=+140.540002997" Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.099988 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:04 crc kubenswrapper[4646]: E1203 10:56:04.100292 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:04.600277248 +0000 UTC m=+141.063333383 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.201087 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:04 crc kubenswrapper[4646]: E1203 10:56:04.201694 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:04.701680123 +0000 UTC m=+141.164736258 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.217067 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ztz2p" podStartSLOduration=122.217034792 podStartE2EDuration="2m2.217034792s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:04.207751357 +0000 UTC m=+140.670807492" watchObservedRunningTime="2025-12-03 10:56:04.217034792 +0000 UTC m=+140.680090927" Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.306939 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:04 crc kubenswrapper[4646]: E1203 10:56:04.307272 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:04.807257968 +0000 UTC m=+141.270314103 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.407930 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:04 crc kubenswrapper[4646]: E1203 10:56:04.408234 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:04.90822462 +0000 UTC m=+141.371280755 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.425682 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.429193 4646 patch_prober.go:28] interesting pod/router-default-5444994796-7whtv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 10:56:04 crc kubenswrapper[4646]: [-]has-synced failed: reason withheld Dec 03 10:56:04 crc kubenswrapper[4646]: [+]process-running ok Dec 03 10:56:04 crc kubenswrapper[4646]: healthz check failed Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.429257 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7whtv" podUID="aa99705d-9e2f-4c32-85cb-04f0cde27411" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.511864 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:04 crc kubenswrapper[4646]: E1203 10:56:04.511982 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:05.011939152 +0000 UTC m=+141.474995287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.512194 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:04 crc kubenswrapper[4646]: E1203 10:56:04.512582 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:05.01257333 +0000 UTC m=+141.475629465 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.612923 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:04 crc kubenswrapper[4646]: E1203 10:56:04.613047 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:05.113023238 +0000 UTC m=+141.576079373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.613243 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:04 crc kubenswrapper[4646]: E1203 10:56:04.613523 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:05.113513762 +0000 UTC m=+141.576569897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.714018 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:04 crc kubenswrapper[4646]: E1203 10:56:04.714294 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:05.214278019 +0000 UTC m=+141.677334154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.814831 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:04 crc kubenswrapper[4646]: E1203 10:56:04.815147 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:05.315134458 +0000 UTC m=+141.778190593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.836204 4646 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-nb6wv container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.836276 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" podUID="7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.917666 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:04 crc kubenswrapper[4646]: E1203 10:56:04.918463 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:05.418443468 +0000 UTC m=+141.881499603 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.936770 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" event={"ID":"53824d3c-1dd1-4f25-b699-de9230490c84","Type":"ContainerStarted","Data":"d2f07402bb1821f6927d42d0aa6ac977a2bf238f7f62b0cd038d2bbde46b80ca"} Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.937380 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.943364 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" event={"ID":"143daef8-3687-4018-8f9f-2a675eac0909","Type":"ContainerStarted","Data":"77faacb42b9bd8c4f04a3464f09b439de0a5af684fb7a1ce01ff4c6a6700e9ae"} Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.946505 4646 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mbfc8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.946556 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" podUID="143daef8-3687-4018-8f9f-2a675eac0909" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.946720 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.969753 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hqnkq" event={"ID":"f9912d55-27ed-44cf-a597-9591fa35476c","Type":"ContainerStarted","Data":"bf7c5c368cbd4090b87b40c40e59e8dce9612719b6dd3784d219b33de740f8d3"} Dec 03 10:56:04 crc kubenswrapper[4646]: I1203 10:56:04.974044 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-prp8k" event={"ID":"75e92964-1833-4889-b175-16f4bba0ed1b","Type":"ContainerStarted","Data":"411dcc08af0cf82813c777f413f045c68c1c9476ac779111de2bf47a5c30a3b1"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.022069 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:05 crc kubenswrapper[4646]: E1203 10:56:05.024950 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:05.524892207 +0000 UTC m=+141.987948342 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.026049 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" event={"ID":"9a0dc011-d040-4ec5-babf-d95b03ec699e","Type":"ContainerStarted","Data":"78812f65e9a705730b131b5167e78ba2f677e13dd64c9dad2b2ecd721e115724"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.048956 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" event={"ID":"cd573c9a-19ea-4975-9645-9d5b9e258d58","Type":"ContainerStarted","Data":"479c56b3b81c4d7e0fc45e0b21f9158257407fd6be415fe662e5c8f63d24fd65"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.086351 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v" event={"ID":"682f390b-3453-40c5-aa8c-5a834ca1d599","Type":"ContainerStarted","Data":"8e494718790c7cba44348595a540af6d5bde7d4a876f567ed65ddcf060d1c47b"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.107621 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" event={"ID":"88293fb3-626a-4944-9d1f-2c569832661d","Type":"ContainerStarted","Data":"f8610e8a476f589818ad54941fe58e6fd9167f0e846d5d16c23a0cb01da1a448"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.108512 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.114423 4646 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-pkrnf container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.114472 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" podUID="88293fb3-626a-4944-9d1f-2c569832661d" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.121741 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6rqmm" event={"ID":"45f218e8-2c99-412f-b85e-956417b5d74a","Type":"ContainerStarted","Data":"fb1700d0d7fc88eaee71c596d0e70273e564cb7539cdf4c64a89e394c61a91a6"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.123146 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:05 crc kubenswrapper[4646]: E1203 10:56:05.123537 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:05.623522583 +0000 UTC m=+142.086578718 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.142487 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" event={"ID":"3cc1272d-e287-4e28-92f7-27dc0aa0d927","Type":"ContainerStarted","Data":"94f2cab92f67d1847b208212ba50439c624ceefb4251361f82b1f1b587da8e9d"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.223996 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p9lsq"] Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.225277 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.226696 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:05 crc kubenswrapper[4646]: E1203 10:56:05.236363 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:05.736328393 +0000 UTC m=+142.199384528 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.237275 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.257985 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p9lsq"] Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.258196 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-4jgxv" event={"ID":"61e9e161-fb19-4da8-9e38-4ca880134e94","Type":"ContainerStarted","Data":"864c888cadb7f1bf4374ff3f27fa7713076c1a288df56a37a572948af7e2aadb"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.303652 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg" event={"ID":"4c5ac722-cd40-44e9-921b-968dbe5e33f1","Type":"ContainerStarted","Data":"e285b7f84fea1da1e91ab690ffeab6a25bdf942e776f7e27ae48d5658ec317a5"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.329633 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.329922 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g62tw\" (UniqueName: \"kubernetes.io/projected/b13d8714-e047-4380-b095-7f5aac84f021-kube-api-access-g62tw\") pod \"community-operators-p9lsq\" (UID: \"b13d8714-e047-4380-b095-7f5aac84f021\") " pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.329957 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b13d8714-e047-4380-b095-7f5aac84f021-utilities\") pod \"community-operators-p9lsq\" (UID: \"b13d8714-e047-4380-b095-7f5aac84f021\") " pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:56:05 crc kubenswrapper[4646]: E1203 10:56:05.330007 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:05.829982197 +0000 UTC m=+142.293038322 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.330078 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b13d8714-e047-4380-b095-7f5aac84f021-catalog-content\") pod \"community-operators-p9lsq\" (UID: \"b13d8714-e047-4380-b095-7f5aac84f021\") " pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.330141 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:05 crc kubenswrapper[4646]: E1203 10:56:05.330512 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:05.830476801 +0000 UTC m=+142.293533056 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.378608 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vkpc9"] Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.380167 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.397922 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.402456 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" event={"ID":"c202a4cb-ebe8-419d-a42d-27a286eeb4a0","Type":"ContainerStarted","Data":"65ef782ff5fcbe9cd509fa116e4b527dc52d1d313a8464d6519b25c9fa55edb7"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.418177 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vkpc9"] Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.419749 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sc5lh" event={"ID":"f20ff0d2-d132-4a30-83b5-e59f6e7260bf","Type":"ContainerStarted","Data":"39e5b7c76c123cfc86a518a2b69330178201c1b3e10f4a7669a7d4e556830e0c"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.432512 4646 patch_prober.go:28] interesting pod/router-default-5444994796-7whtv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 10:56:05 crc kubenswrapper[4646]: [-]has-synced failed: reason withheld Dec 03 10:56:05 crc kubenswrapper[4646]: [+]process-running ok Dec 03 10:56:05 crc kubenswrapper[4646]: healthz check failed Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.432557 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7whtv" podUID="aa99705d-9e2f-4c32-85cb-04f0cde27411" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.432915 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.433127 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1141f44-4444-47fb-878e-7a17db767b8d-utilities\") pod \"certified-operators-vkpc9\" (UID: \"d1141f44-4444-47fb-878e-7a17db767b8d\") " pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.433166 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1141f44-4444-47fb-878e-7a17db767b8d-catalog-content\") pod \"certified-operators-vkpc9\" (UID: \"d1141f44-4444-47fb-878e-7a17db767b8d\") " pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.433190 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g62tw\" (UniqueName: \"kubernetes.io/projected/b13d8714-e047-4380-b095-7f5aac84f021-kube-api-access-g62tw\") pod \"community-operators-p9lsq\" (UID: \"b13d8714-e047-4380-b095-7f5aac84f021\") " pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.433216 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b13d8714-e047-4380-b095-7f5aac84f021-utilities\") pod \"community-operators-p9lsq\" (UID: \"b13d8714-e047-4380-b095-7f5aac84f021\") " pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.433282 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b13d8714-e047-4380-b095-7f5aac84f021-catalog-content\") pod \"community-operators-p9lsq\" (UID: \"b13d8714-e047-4380-b095-7f5aac84f021\") " pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.433352 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts5nb\" (UniqueName: \"kubernetes.io/projected/d1141f44-4444-47fb-878e-7a17db767b8d-kube-api-access-ts5nb\") pod \"certified-operators-vkpc9\" (UID: \"d1141f44-4444-47fb-878e-7a17db767b8d\") " pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:56:05 crc kubenswrapper[4646]: E1203 10:56:05.433470 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:05.933456061 +0000 UTC m=+142.396512196 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.434504 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b13d8714-e047-4380-b095-7f5aac84f021-utilities\") pod \"community-operators-p9lsq\" (UID: \"b13d8714-e047-4380-b095-7f5aac84f021\") " pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.435163 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b13d8714-e047-4380-b095-7f5aac84f021-catalog-content\") pod \"community-operators-p9lsq\" (UID: \"b13d8714-e047-4380-b095-7f5aac84f021\") " pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.464987 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj" event={"ID":"341c245b-3af9-42fd-9592-00737bcc7ba7","Type":"ContainerStarted","Data":"7bf7386f77311def64225ce62a55310a926dc2f2b147d4f9d5f98c66e9f393a4"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.489228 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-vtndh" event={"ID":"40507d60-aae3-43da-85b0-09c02327ece8","Type":"ContainerStarted","Data":"7b68b6734948455342a3dc309d6f3a1a0ef75c7a5845e31def516e3ff5cf3314"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.489270 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-vtndh" event={"ID":"40507d60-aae3-43da-85b0-09c02327ece8","Type":"ContainerStarted","Data":"7bb9c37000268c927c527c2c72dcc833bf3cda8d8028eee05d4c1f92c3a7e080"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.528650 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" event={"ID":"9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450","Type":"ContainerStarted","Data":"166df0b8dce64af128bfb8759f0dceefadaaccbfe127ea575f91f7e5608af505"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.528711 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" event={"ID":"9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450","Type":"ContainerStarted","Data":"5b16e05fcdf856fb244ab2204658df50fe433f158f0966288ed5ab2403ea76ba"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.534512 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts5nb\" (UniqueName: \"kubernetes.io/projected/d1141f44-4444-47fb-878e-7a17db767b8d-kube-api-access-ts5nb\") pod \"certified-operators-vkpc9\" (UID: \"d1141f44-4444-47fb-878e-7a17db767b8d\") " pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.534690 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.534818 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1141f44-4444-47fb-878e-7a17db767b8d-utilities\") pod \"certified-operators-vkpc9\" (UID: \"d1141f44-4444-47fb-878e-7a17db767b8d\") " pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.534964 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1141f44-4444-47fb-878e-7a17db767b8d-catalog-content\") pod \"certified-operators-vkpc9\" (UID: \"d1141f44-4444-47fb-878e-7a17db767b8d\") " pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:56:05 crc kubenswrapper[4646]: E1203 10:56:05.535241 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:06.035227667 +0000 UTC m=+142.498283802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.535801 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1141f44-4444-47fb-878e-7a17db767b8d-catalog-content\") pod \"certified-operators-vkpc9\" (UID: \"d1141f44-4444-47fb-878e-7a17db767b8d\") " pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.536426 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1141f44-4444-47fb-878e-7a17db767b8d-utilities\") pod \"certified-operators-vkpc9\" (UID: \"d1141f44-4444-47fb-878e-7a17db767b8d\") " pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.549233 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd" event={"ID":"464c6029-4dac-48ee-9f4f-70659bb6445b","Type":"ContainerStarted","Data":"3f4a8d7a29ef80b58a9f100dc19bb061eae2d2dd2412f5b7be417c63a15ed911"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.558058 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm" event={"ID":"9dc04fc6-6a79-4317-8b14-be963345a26c","Type":"ContainerStarted","Data":"473133a41a32efcab8a63d4315d992afb37980cf91bf9a40cf77cc5bd22823b2"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.576693 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kv72s"] Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.578467 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.591265 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g62tw\" (UniqueName: \"kubernetes.io/projected/b13d8714-e047-4380-b095-7f5aac84f021-kube-api-access-g62tw\") pod \"community-operators-p9lsq\" (UID: \"b13d8714-e047-4380-b095-7f5aac84f021\") " pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.596555 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" event={"ID":"500b7283-f949-4ff4-a868-024bc3a8a6e5","Type":"ContainerStarted","Data":"5f168deea9864c2c1bdc6f852b4a6b290cc8e2316a7e389f18624ab2cc004c44"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.604973 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts5nb\" (UniqueName: \"kubernetes.io/projected/d1141f44-4444-47fb-878e-7a17db767b8d-kube-api-access-ts5nb\") pod \"certified-operators-vkpc9\" (UID: \"d1141f44-4444-47fb-878e-7a17db767b8d\") " pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.625897 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kv72s"] Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.627876 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-55nsv" event={"ID":"041d1b1a-4861-4b7e-98ee-f5707404c173","Type":"ContainerStarted","Data":"54e3afdb68ba5d97b5e94450578807d942398af28110bb30775db0426c1780fa"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.655168 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" podStartSLOduration=123.655145201 podStartE2EDuration="2m3.655145201s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:05.635828149 +0000 UTC m=+142.098884304" watchObservedRunningTime="2025-12-03 10:56:05.655145201 +0000 UTC m=+142.118201336" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.658088 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.638181 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.659550 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxktn\" (UniqueName: \"kubernetes.io/projected/2b646c66-7457-42fe-abe5-0def6d0d3262-kube-api-access-rxktn\") pod \"community-operators-kv72s\" (UID: \"2b646c66-7457-42fe-abe5-0def6d0d3262\") " pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.660449 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b646c66-7457-42fe-abe5-0def6d0d3262-utilities\") pod \"community-operators-kv72s\" (UID: \"2b646c66-7457-42fe-abe5-0def6d0d3262\") " pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.660899 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b646c66-7457-42fe-abe5-0def6d0d3262-catalog-content\") pod \"community-operators-kv72s\" (UID: \"2b646c66-7457-42fe-abe5-0def6d0d3262\") " pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:56:05 crc kubenswrapper[4646]: E1203 10:56:05.666342 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:06.1663117 +0000 UTC m=+142.629367835 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.697476 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-4jgxv" podStartSLOduration=123.697459119 podStartE2EDuration="2m3.697459119s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:05.693954839 +0000 UTC m=+142.157010974" watchObservedRunningTime="2025-12-03 10:56:05.697459119 +0000 UTC m=+142.160515254" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.741366 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5wscd" event={"ID":"697c16ea-e85a-4b7c-97f2-1d2a9ff7b5ed","Type":"ContainerStarted","Data":"f3ae62569e179014e530542840b33be38af37368e8df176bc7815f8675a125eb"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.754290 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.763065 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.763363 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b646c66-7457-42fe-abe5-0def6d0d3262-utilities\") pod \"community-operators-kv72s\" (UID: \"2b646c66-7457-42fe-abe5-0def6d0d3262\") " pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.763570 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b646c66-7457-42fe-abe5-0def6d0d3262-catalog-content\") pod \"community-operators-kv72s\" (UID: \"2b646c66-7457-42fe-abe5-0def6d0d3262\") " pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.763673 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxktn\" (UniqueName: \"kubernetes.io/projected/2b646c66-7457-42fe-abe5-0def6d0d3262-kube-api-access-rxktn\") pod \"community-operators-kv72s\" (UID: \"2b646c66-7457-42fe-abe5-0def6d0d3262\") " pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.764412 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b646c66-7457-42fe-abe5-0def6d0d3262-utilities\") pod \"community-operators-kv72s\" (UID: \"2b646c66-7457-42fe-abe5-0def6d0d3262\") " pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.764586 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hc78d" event={"ID":"92576279-7206-4601-9734-85243e90b5ed","Type":"ContainerStarted","Data":"7af3d70abe5dda2c545d595cf7dc08aacdae80043086c4a4ed878558cc5d6885"} Dec 03 10:56:05 crc kubenswrapper[4646]: E1203 10:56:05.764846 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:06.264833542 +0000 UTC m=+142.727889677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.765239 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b646c66-7457-42fe-abe5-0def6d0d3262-catalog-content\") pod \"community-operators-kv72s\" (UID: \"2b646c66-7457-42fe-abe5-0def6d0d3262\") " pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.804680 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v" event={"ID":"578c8a44-77e3-4c21-b841-3fc93a82b0f1","Type":"ContainerStarted","Data":"54a4b50c8ac6e608d73ab30f369eef3fde28fc986ff2223daa04d975587df9d3"} Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.808808 4646 patch_prober.go:28] interesting pod/downloads-7954f5f757-l2kck container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.808859 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-l2kck" podUID="54c0778a-7e52-4da6-ab86-6a3492eb278b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.809292 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" podStartSLOduration=123.809270761 podStartE2EDuration="2m3.809270761s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:05.808688134 +0000 UTC m=+142.271744269" watchObservedRunningTime="2025-12-03 10:56:05.809270761 +0000 UTC m=+142.272326896" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.811461 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxktn\" (UniqueName: \"kubernetes.io/projected/2b646c66-7457-42fe-abe5-0def6d0d3262-kube-api-access-rxktn\") pod \"community-operators-kv72s\" (UID: \"2b646c66-7457-42fe-abe5-0def6d0d3262\") " pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.823489 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rzw7n"] Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.824659 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.865209 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:05 crc kubenswrapper[4646]: E1203 10:56:05.867207 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:06.367182765 +0000 UTC m=+142.830238900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.896996 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rzw7n"] Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.928619 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.967645 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92bfb497-1e60-4973-a76c-0c686689cc53-catalog-content\") pod \"certified-operators-rzw7n\" (UID: \"92bfb497-1e60-4973-a76c-0c686689cc53\") " pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.967803 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.967929 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92bfb497-1e60-4973-a76c-0c686689cc53-utilities\") pod \"certified-operators-rzw7n\" (UID: \"92bfb497-1e60-4973-a76c-0c686689cc53\") " pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.968010 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6vcv\" (UniqueName: \"kubernetes.io/projected/92bfb497-1e60-4973-a76c-0c686689cc53-kube-api-access-s6vcv\") pod \"certified-operators-rzw7n\" (UID: \"92bfb497-1e60-4973-a76c-0c686689cc53\") " pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:56:05 crc kubenswrapper[4646]: E1203 10:56:05.968358 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:06.468327912 +0000 UTC m=+142.931384047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.992837 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" podStartSLOduration=123.992817812 podStartE2EDuration="2m3.992817812s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:05.941026283 +0000 UTC m=+142.404082418" watchObservedRunningTime="2025-12-03 10:56:05.992817812 +0000 UTC m=+142.455873947" Dec 03 10:56:05 crc kubenswrapper[4646]: I1203 10:56:05.995286 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-vtndh" podStartSLOduration=8.995280522 podStartE2EDuration="8.995280522s" podCreationTimestamp="2025-12-03 10:55:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:05.994261383 +0000 UTC m=+142.457317518" watchObservedRunningTime="2025-12-03 10:56:05.995280522 +0000 UTC m=+142.458336657" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.069295 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.069811 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92bfb497-1e60-4973-a76c-0c686689cc53-catalog-content\") pod \"certified-operators-rzw7n\" (UID: \"92bfb497-1e60-4973-a76c-0c686689cc53\") " pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.069953 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92bfb497-1e60-4973-a76c-0c686689cc53-utilities\") pod \"certified-operators-rzw7n\" (UID: \"92bfb497-1e60-4973-a76c-0c686689cc53\") " pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.069983 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6vcv\" (UniqueName: \"kubernetes.io/projected/92bfb497-1e60-4973-a76c-0c686689cc53-kube-api-access-s6vcv\") pod \"certified-operators-rzw7n\" (UID: \"92bfb497-1e60-4973-a76c-0c686689cc53\") " pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.070752 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92bfb497-1e60-4973-a76c-0c686689cc53-utilities\") pod \"certified-operators-rzw7n\" (UID: \"92bfb497-1e60-4973-a76c-0c686689cc53\") " pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.071092 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92bfb497-1e60-4973-a76c-0c686689cc53-catalog-content\") pod \"certified-operators-rzw7n\" (UID: \"92bfb497-1e60-4973-a76c-0c686689cc53\") " pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:56:06 crc kubenswrapper[4646]: E1203 10:56:06.071175 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:06.571160658 +0000 UTC m=+143.034216793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.090966 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sc5lh" podStartSLOduration=124.090946223 podStartE2EDuration="2m4.090946223s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:06.056978663 +0000 UTC m=+142.520034798" watchObservedRunningTime="2025-12-03 10:56:06.090946223 +0000 UTC m=+142.554002358" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.125423 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cb8tj" podStartSLOduration=124.125407207 podStartE2EDuration="2m4.125407207s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:06.091065317 +0000 UTC m=+142.554121452" watchObservedRunningTime="2025-12-03 10:56:06.125407207 +0000 UTC m=+142.588463342" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.129259 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6vcv\" (UniqueName: \"kubernetes.io/projected/92bfb497-1e60-4973-a76c-0c686689cc53-kube-api-access-s6vcv\") pod \"certified-operators-rzw7n\" (UID: \"92bfb497-1e60-4973-a76c-0c686689cc53\") " pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.171960 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:06 crc kubenswrapper[4646]: E1203 10:56:06.172521 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:06.672510282 +0000 UTC m=+143.135566417 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.242303 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.272007 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" podStartSLOduration=124.271992172 podStartE2EDuration="2m4.271992172s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:06.195391355 +0000 UTC m=+142.658447490" watchObservedRunningTime="2025-12-03 10:56:06.271992172 +0000 UTC m=+142.735048307" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.273846 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:06 crc kubenswrapper[4646]: E1203 10:56:06.274187 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:06.774174985 +0000 UTC m=+143.237231120 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.298211 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wndxj" podStartSLOduration=124.29819469 podStartE2EDuration="2m4.29819469s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:06.273424483 +0000 UTC m=+142.736480618" watchObservedRunningTime="2025-12-03 10:56:06.29819469 +0000 UTC m=+142.761250825" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.320798 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.375825 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:06 crc kubenswrapper[4646]: E1203 10:56:06.376166 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:06.876153006 +0000 UTC m=+143.339209141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.435981 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-55nsv" podStartSLOduration=124.435965664 podStartE2EDuration="2m4.435965664s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:06.389304012 +0000 UTC m=+142.852360147" watchObservedRunningTime="2025-12-03 10:56:06.435965664 +0000 UTC m=+142.899021799" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.445637 4646 patch_prober.go:28] interesting pod/router-default-5444994796-7whtv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 10:56:06 crc kubenswrapper[4646]: [-]has-synced failed: reason withheld Dec 03 10:56:06 crc kubenswrapper[4646]: [+]process-running ok Dec 03 10:56:06 crc kubenswrapper[4646]: healthz check failed Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.445933 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7whtv" podUID="aa99705d-9e2f-4c32-85cb-04f0cde27411" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.477536 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:06 crc kubenswrapper[4646]: E1203 10:56:06.477656 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:06.977632264 +0000 UTC m=+143.440688399 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.478721 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:06 crc kubenswrapper[4646]: E1203 10:56:06.479063 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:06.979054294 +0000 UTC m=+143.442110429 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.580524 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:06 crc kubenswrapper[4646]: E1203 10:56:06.580689 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:07.080672485 +0000 UTC m=+143.543728620 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.580730 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:06 crc kubenswrapper[4646]: E1203 10:56:06.581071 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:07.081056246 +0000 UTC m=+143.544112381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.683262 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:06 crc kubenswrapper[4646]: E1203 10:56:06.683684 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:07.183665216 +0000 UTC m=+143.646721341 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.787611 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:06 crc kubenswrapper[4646]: E1203 10:56:06.787911 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:07.287900782 +0000 UTC m=+143.750956917 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.822454 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hqnkq" event={"ID":"f9912d55-27ed-44cf-a597-9591fa35476c","Type":"ContainerStarted","Data":"12edcd0bc2c00c93ac52e8fe66e59c2163a86e7197d311b3c60f61359aeda349"} Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.831188 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm" event={"ID":"9dc04fc6-6a79-4317-8b14-be963345a26c","Type":"ContainerStarted","Data":"0df1d3bbdc102a85fa78a8113e501f3a97d5aa000a24989f511d4bc82fd69e9b"} Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.844945 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-prp8k" event={"ID":"75e92964-1833-4889-b175-16f4bba0ed1b","Type":"ContainerStarted","Data":"547505b8a22fdd90566671cd1ec7da8fb91c312166dacea8e4f42ab261db9138"} Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.845003 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-prp8k" event={"ID":"75e92964-1833-4889-b175-16f4bba0ed1b","Type":"ContainerStarted","Data":"b0ffb5b40025da58b34611159dd1f5f61ec80474f7adba4402d59595e3dc76b5"} Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.849243 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cg4cm" podStartSLOduration=124.849216553 podStartE2EDuration="2m4.849216553s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:06.848191533 +0000 UTC m=+143.311247668" watchObservedRunningTime="2025-12-03 10:56:06.849216553 +0000 UTC m=+143.312272688" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.850305 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6rqmm" event={"ID":"45f218e8-2c99-412f-b85e-956417b5d74a","Type":"ContainerStarted","Data":"c27d0254c211dda526ebf509ea7e92367a00d1d632fc7e38a2f5e18b59509c63"} Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.867304 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd" event={"ID":"464c6029-4dac-48ee-9f4f-70659bb6445b","Type":"ContainerStarted","Data":"0fbcb0c49f34ce572e198f445d42cff39fdb9c56f842f7381d9657973ee048dc"} Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.887638 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hc78d" event={"ID":"92576279-7206-4601-9734-85243e90b5ed","Type":"ContainerStarted","Data":"97237f71a2fdff06e191ad964a836f3bcdb7781ab046696c279d541be1aa394c"} Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.888434 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-hc78d" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.888788 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:06 crc kubenswrapper[4646]: E1203 10:56:06.889874 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:07.389859553 +0000 UTC m=+143.852915688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.904536 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v" event={"ID":"682f390b-3453-40c5-aa8c-5a834ca1d599","Type":"ContainerStarted","Data":"bf507c0c0f219eacd2c05206b7a675db41eb6680e1dbd42d4f79990405bd204e"} Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.918755 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg" event={"ID":"4c5ac722-cd40-44e9-921b-968dbe5e33f1","Type":"ContainerStarted","Data":"283f7d76ba40fafdd84e583c2197409ce9fec6981ebd87c2228ad5d0bf9dc0b1"} Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.924527 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-prp8k" podStartSLOduration=124.924511812 podStartE2EDuration="2m4.924511812s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:06.877124959 +0000 UTC m=+143.340181094" watchObservedRunningTime="2025-12-03 10:56:06.924511812 +0000 UTC m=+143.387567947" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.927541 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzd" podStartSLOduration=124.927533488 podStartE2EDuration="2m4.927533488s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:06.926021265 +0000 UTC m=+143.389077400" watchObservedRunningTime="2025-12-03 10:56:06.927533488 +0000 UTC m=+143.390589623" Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.990875 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-czsqx" event={"ID":"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6","Type":"ContainerStarted","Data":"618757eb2243858d687058bca19cb18c6f95750ea43507fbdc228604f286870d"} Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.992036 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:06 crc kubenswrapper[4646]: E1203 10:56:06.993895 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:07.493882973 +0000 UTC m=+143.956939108 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:06 crc kubenswrapper[4646]: I1203 10:56:06.999363 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-6rqmm" podStartSLOduration=123.999318868 podStartE2EDuration="2m3.999318868s" podCreationTimestamp="2025-12-03 10:54:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:06.95384453 +0000 UTC m=+143.416900665" watchObservedRunningTime="2025-12-03 10:56:06.999318868 +0000 UTC m=+143.462375003" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.033555 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-hc78d" podStartSLOduration=10.033535335 podStartE2EDuration="10.033535335s" podCreationTimestamp="2025-12-03 10:55:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:07.004801014 +0000 UTC m=+143.467857149" watchObservedRunningTime="2025-12-03 10:56:07.033535335 +0000 UTC m=+143.496591470" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.035918 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wgpqg" podStartSLOduration=124.035905553 podStartE2EDuration="2m4.035905553s" podCreationTimestamp="2025-12-03 10:54:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:07.031638571 +0000 UTC m=+143.494694706" watchObservedRunningTime="2025-12-03 10:56:07.035905553 +0000 UTC m=+143.498961698" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.040596 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" event={"ID":"cd573c9a-19ea-4975-9645-9d5b9e258d58","Type":"ContainerStarted","Data":"10e5a6c92acc7cd8d0aedbcdfae764f3eb2c9fd121726b2c647bca8bc33a172a"} Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.041383 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.083625 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" event={"ID":"324debd2-da21-490a-9e83-36fbcb470aab","Type":"ContainerStarted","Data":"b80a9a2b65d6bdb2c1888fa30acef643b41e0c46085f57a0ef387a266a0ff52f"} Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.098930 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5dz9v" podStartSLOduration=125.098915612 podStartE2EDuration="2m5.098915612s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:07.095506834 +0000 UTC m=+143.558562969" watchObservedRunningTime="2025-12-03 10:56:07.098915612 +0000 UTC m=+143.561971747" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.101986 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:07 crc kubenswrapper[4646]: E1203 10:56:07.102928 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:07.602915766 +0000 UTC m=+144.065971901 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.130969 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v" event={"ID":"578c8a44-77e3-4c21-b841-3fc93a82b0f1","Type":"ContainerStarted","Data":"194648c083db8cdabd7211496dd393ce1c24e431aa02ac102acd87a5cf647625"} Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.131003 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.143752 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" event={"ID":"9a0dc011-d040-4ec5-babf-d95b03ec699e","Type":"ContainerStarted","Data":"4d32704462507aa4f2b6cca952cb02673970cb84b880fa132b2d60c9cc1b8513"} Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.154620 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" event={"ID":"3cc1272d-e287-4e28-92f7-27dc0aa0d927","Type":"ContainerStarted","Data":"2a16ec9b80875040a0678b34498fc7df1b9ac97e5179922643c8a2dd6c79a5e8"} Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.168974 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-52ndk" event={"ID":"e6da1553-a91a-4055-af3d-bd30b20e6c0f","Type":"ContainerStarted","Data":"dc5bc70e262493e5ff27f02b5d2f4995c359404d63b8d7193381cdf2f81ea92e"} Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.170239 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.194104 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" event={"ID":"9af4fbd9-cad9-4d67-9ee7-b3ff1fb0c450","Type":"ContainerStarted","Data":"c649477c5c723b94a0b0003784e7b7800070dfbef6c17f2191bb5373ab886081"} Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.199207 4646 patch_prober.go:28] interesting pod/downloads-7954f5f757-l2kck container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.199238 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-l2kck" podUID="54c0778a-7e52-4da6-ab86-6a3492eb278b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.199727 4646 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mbfc8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.199746 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" podUID="143daef8-3687-4018-8f9f-2a675eac0909" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.201849 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p9lsq"] Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.204040 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:07 crc kubenswrapper[4646]: E1203 10:56:07.206573 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:07.706563334 +0000 UTC m=+144.169619469 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.230734 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ntvnm" podStartSLOduration=125.230717604 podStartE2EDuration="2m5.230717604s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:07.230072725 +0000 UTC m=+143.693128860" watchObservedRunningTime="2025-12-03 10:56:07.230717604 +0000 UTC m=+143.693773739" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.237552 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pkrnf" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.254487 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-697vl" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.274035 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vkpc9"] Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.305510 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kv72s"] Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.306756 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:07 crc kubenswrapper[4646]: E1203 10:56:07.308320 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:07.808297209 +0000 UTC m=+144.271353344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.329374 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:07 crc kubenswrapper[4646]: E1203 10:56:07.330636 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:07.830619656 +0000 UTC m=+144.293675791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.413033 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" podStartSLOduration=125.413015398 podStartE2EDuration="2m5.413015398s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:07.367584231 +0000 UTC m=+143.830640366" watchObservedRunningTime="2025-12-03 10:56:07.413015398 +0000 UTC m=+143.876071523" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.415474 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xjl9g"] Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.430900 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:07 crc kubenswrapper[4646]: E1203 10:56:07.431219 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:07.931204498 +0000 UTC m=+144.394260633 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.435647 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.451495 4646 patch_prober.go:28] interesting pod/router-default-5444994796-7whtv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 10:56:07 crc kubenswrapper[4646]: [-]has-synced failed: reason withheld Dec 03 10:56:07 crc kubenswrapper[4646]: [+]process-running ok Dec 03 10:56:07 crc kubenswrapper[4646]: healthz check failed Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.451542 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7whtv" podUID="aa99705d-9e2f-4c32-85cb-04f0cde27411" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.462830 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.539178 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.539247 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d20d9814-aa42-4109-9740-d0e5e6205a00-utilities\") pod \"redhat-marketplace-xjl9g\" (UID: \"d20d9814-aa42-4109-9740-d0e5e6205a00\") " pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.539273 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxd62\" (UniqueName: \"kubernetes.io/projected/d20d9814-aa42-4109-9740-d0e5e6205a00-kube-api-access-xxd62\") pod \"redhat-marketplace-xjl9g\" (UID: \"d20d9814-aa42-4109-9740-d0e5e6205a00\") " pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.539415 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d20d9814-aa42-4109-9740-d0e5e6205a00-catalog-content\") pod \"redhat-marketplace-xjl9g\" (UID: \"d20d9814-aa42-4109-9740-d0e5e6205a00\") " pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:56:07 crc kubenswrapper[4646]: E1203 10:56:07.539676 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:08.039663924 +0000 UTC m=+144.502720059 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.561912 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xjl9g"] Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.595639 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rzw7n"] Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.619106 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-52ndk" podStartSLOduration=125.619093082 podStartE2EDuration="2m5.619093082s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:07.617325311 +0000 UTC m=+144.080381446" watchObservedRunningTime="2025-12-03 10:56:07.619093082 +0000 UTC m=+144.082149217" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.644842 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.645049 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d20d9814-aa42-4109-9740-d0e5e6205a00-catalog-content\") pod \"redhat-marketplace-xjl9g\" (UID: \"d20d9814-aa42-4109-9740-d0e5e6205a00\") " pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.645141 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d20d9814-aa42-4109-9740-d0e5e6205a00-utilities\") pod \"redhat-marketplace-xjl9g\" (UID: \"d20d9814-aa42-4109-9740-d0e5e6205a00\") " pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.645188 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxd62\" (UniqueName: \"kubernetes.io/projected/d20d9814-aa42-4109-9740-d0e5e6205a00-kube-api-access-xxd62\") pod \"redhat-marketplace-xjl9g\" (UID: \"d20d9814-aa42-4109-9740-d0e5e6205a00\") " pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:56:07 crc kubenswrapper[4646]: E1203 10:56:07.645621 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:08.145604719 +0000 UTC m=+144.608660854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.645961 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d20d9814-aa42-4109-9740-d0e5e6205a00-catalog-content\") pod \"redhat-marketplace-xjl9g\" (UID: \"d20d9814-aa42-4109-9740-d0e5e6205a00\") " pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.646166 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d20d9814-aa42-4109-9740-d0e5e6205a00-utilities\") pod \"redhat-marketplace-xjl9g\" (UID: \"d20d9814-aa42-4109-9740-d0e5e6205a00\") " pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.712768 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxd62\" (UniqueName: \"kubernetes.io/projected/d20d9814-aa42-4109-9740-d0e5e6205a00-kube-api-access-xxd62\") pod \"redhat-marketplace-xjl9g\" (UID: \"d20d9814-aa42-4109-9740-d0e5e6205a00\") " pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.721982 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v" podStartSLOduration=125.721950879 podStartE2EDuration="2m5.721950879s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:07.720487017 +0000 UTC m=+144.183543152" watchObservedRunningTime="2025-12-03 10:56:07.721950879 +0000 UTC m=+144.185007014" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.746120 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:07 crc kubenswrapper[4646]: E1203 10:56:07.746473 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:08.246461438 +0000 UTC m=+144.709517573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.768267 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ft6s8"] Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.769107 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.769447 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.805911 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ft6s8"] Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.863260 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.863667 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4048178-b02e-472c-a4db-5520e198217c-utilities\") pod \"redhat-marketplace-ft6s8\" (UID: \"d4048178-b02e-472c-a4db-5520e198217c\") " pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.863693 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4048178-b02e-472c-a4db-5520e198217c-catalog-content\") pod \"redhat-marketplace-ft6s8\" (UID: \"d4048178-b02e-472c-a4db-5520e198217c\") " pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.863809 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crlwl\" (UniqueName: \"kubernetes.io/projected/d4048178-b02e-472c-a4db-5520e198217c-kube-api-access-crlwl\") pod \"redhat-marketplace-ft6s8\" (UID: \"d4048178-b02e-472c-a4db-5520e198217c\") " pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:56:07 crc kubenswrapper[4646]: E1203 10:56:07.863977 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:08.363961743 +0000 UTC m=+144.827017878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.882066 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lkssm" podStartSLOduration=125.882047119 podStartE2EDuration="2m5.882047119s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:07.806019719 +0000 UTC m=+144.269075854" watchObservedRunningTime="2025-12-03 10:56:07.882047119 +0000 UTC m=+144.345103254" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.882923 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" podStartSLOduration=125.882917394 podStartE2EDuration="2m5.882917394s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:07.880420963 +0000 UTC m=+144.343477098" watchObservedRunningTime="2025-12-03 10:56:07.882917394 +0000 UTC m=+144.345973529" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.956826 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-4ftgs" podStartSLOduration=125.956796253 podStartE2EDuration="2m5.956796253s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:07.95493511 +0000 UTC m=+144.417991245" watchObservedRunningTime="2025-12-03 10:56:07.956796253 +0000 UTC m=+144.419852388" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.964959 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4048178-b02e-472c-a4db-5520e198217c-utilities\") pod \"redhat-marketplace-ft6s8\" (UID: \"d4048178-b02e-472c-a4db-5520e198217c\") " pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.965007 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4048178-b02e-472c-a4db-5520e198217c-catalog-content\") pod \"redhat-marketplace-ft6s8\" (UID: \"d4048178-b02e-472c-a4db-5520e198217c\") " pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.965054 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.965100 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crlwl\" (UniqueName: \"kubernetes.io/projected/d4048178-b02e-472c-a4db-5520e198217c-kube-api-access-crlwl\") pod \"redhat-marketplace-ft6s8\" (UID: \"d4048178-b02e-472c-a4db-5520e198217c\") " pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.966325 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4048178-b02e-472c-a4db-5520e198217c-utilities\") pod \"redhat-marketplace-ft6s8\" (UID: \"d4048178-b02e-472c-a4db-5520e198217c\") " pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.966639 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4048178-b02e-472c-a4db-5520e198217c-catalog-content\") pod \"redhat-marketplace-ft6s8\" (UID: \"d4048178-b02e-472c-a4db-5520e198217c\") " pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:56:07 crc kubenswrapper[4646]: E1203 10:56:07.966918 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:08.466905402 +0000 UTC m=+144.929961537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:07 crc kubenswrapper[4646]: I1203 10:56:07.986638 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crlwl\" (UniqueName: \"kubernetes.io/projected/d4048178-b02e-472c-a4db-5520e198217c-kube-api-access-crlwl\") pod \"redhat-marketplace-ft6s8\" (UID: \"d4048178-b02e-472c-a4db-5520e198217c\") " pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.042286 4646 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-9x4fj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.042356 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" podUID="cd573c9a-19ea-4975-9645-9d5b9e258d58" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.065811 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:08 crc kubenswrapper[4646]: E1203 10:56:08.066281 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:08.566233028 +0000 UTC m=+145.029289163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.170500 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:08 crc kubenswrapper[4646]: E1203 10:56:08.170834 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:08.670821424 +0000 UTC m=+145.133877559 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.171694 4646 patch_prober.go:28] interesting pod/console-operator-58897d9998-52ndk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.171753 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-52ndk" podUID="e6da1553-a91a-4055-af3d-bd30b20e6c0f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.17:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.262724 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p9lsq" event={"ID":"b13d8714-e047-4380-b095-7f5aac84f021","Type":"ContainerStarted","Data":"7f71624badfa9a269a1e00b1617569675cad4d6c15b7a3bef791825b406ea716"} Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.263670 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.273026 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:08 crc kubenswrapper[4646]: E1203 10:56:08.273375 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:08.773359371 +0000 UTC m=+145.236415506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.281068 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzw7n" event={"ID":"92bfb497-1e60-4973-a76c-0c686689cc53","Type":"ContainerStarted","Data":"134b6d27fa65c504636837c4441c8d5eb488079d070ffe838b86441d2e24dcd6"} Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.282438 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-czsqx" event={"ID":"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6","Type":"ContainerStarted","Data":"a48f0a4c1a303d884d4dbe24cd25ebd4932bc3cd3d8fbafdf5e7ed02a233092f"} Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.283475 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkpc9" event={"ID":"d1141f44-4444-47fb-878e-7a17db767b8d","Type":"ContainerStarted","Data":"fd0584fef9cbb97cae3b97f8c6905ac586fb2de2cfd684fa21c8a5babd791f44"} Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.300514 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kv72s" event={"ID":"2b646c66-7457-42fe-abe5-0def6d0d3262","Type":"ContainerStarted","Data":"eed51c0a4ca3ceeed1093d11420d8f781db8e0d56b0cff6dedab7d72141e87ae"} Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.317056 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hqnkq" event={"ID":"f9912d55-27ed-44cf-a597-9591fa35476c","Type":"ContainerStarted","Data":"316e5252583473feba17f43e3dedd7676a0db8fd02ac5be251cfe0847154b144"} Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.318552 4646 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mbfc8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.318585 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" podUID="143daef8-3687-4018-8f9f-2a675eac0909" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.375971 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:08 crc kubenswrapper[4646]: E1203 10:56:08.376418 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:08.876404974 +0000 UTC m=+145.339461109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.433571 4646 patch_prober.go:28] interesting pod/router-default-5444994796-7whtv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 10:56:08 crc kubenswrapper[4646]: [-]has-synced failed: reason withheld Dec 03 10:56:08 crc kubenswrapper[4646]: [+]process-running ok Dec 03 10:56:08 crc kubenswrapper[4646]: healthz check failed Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.433638 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7whtv" podUID="aa99705d-9e2f-4c32-85cb-04f0cde27411" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.434451 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-hqnkq" podStartSLOduration=126.4344328 podStartE2EDuration="2m6.4344328s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:08.422935882 +0000 UTC m=+144.885992017" watchObservedRunningTime="2025-12-03 10:56:08.4344328 +0000 UTC m=+144.897488935" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.437090 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xjl9g"] Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.479477 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:08 crc kubenswrapper[4646]: E1203 10:56:08.481165 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:08.981143934 +0000 UTC m=+145.444200069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.568481 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vsmhw"] Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.569484 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.572924 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.581373 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:08 crc kubenswrapper[4646]: E1203 10:56:08.581717 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:09.081704215 +0000 UTC m=+145.544760350 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.590835 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vsmhw"] Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.683300 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.683888 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/250f331c-acf3-4b28-b2a4-515a8cceb844-utilities\") pod \"redhat-operators-vsmhw\" (UID: \"250f331c-acf3-4b28-b2a4-515a8cceb844\") " pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.683932 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6vrt\" (UniqueName: \"kubernetes.io/projected/250f331c-acf3-4b28-b2a4-515a8cceb844-kube-api-access-m6vrt\") pod \"redhat-operators-vsmhw\" (UID: \"250f331c-acf3-4b28-b2a4-515a8cceb844\") " pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:56:08 crc kubenswrapper[4646]: E1203 10:56:08.683983 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:09.183961014 +0000 UTC m=+145.647017149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.684103 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/250f331c-acf3-4b28-b2a4-515a8cceb844-catalog-content\") pod \"redhat-operators-vsmhw\" (UID: \"250f331c-acf3-4b28-b2a4-515a8cceb844\") " pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.787977 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6vrt\" (UniqueName: \"kubernetes.io/projected/250f331c-acf3-4b28-b2a4-515a8cceb844-kube-api-access-m6vrt\") pod \"redhat-operators-vsmhw\" (UID: \"250f331c-acf3-4b28-b2a4-515a8cceb844\") " pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.788256 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/250f331c-acf3-4b28-b2a4-515a8cceb844-catalog-content\") pod \"redhat-operators-vsmhw\" (UID: \"250f331c-acf3-4b28-b2a4-515a8cceb844\") " pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.788317 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.788381 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/250f331c-acf3-4b28-b2a4-515a8cceb844-utilities\") pod \"redhat-operators-vsmhw\" (UID: \"250f331c-acf3-4b28-b2a4-515a8cceb844\") " pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.788770 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/250f331c-acf3-4b28-b2a4-515a8cceb844-utilities\") pod \"redhat-operators-vsmhw\" (UID: \"250f331c-acf3-4b28-b2a4-515a8cceb844\") " pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:56:08 crc kubenswrapper[4646]: E1203 10:56:08.789008 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:09.288996753 +0000 UTC m=+145.752052888 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.789003 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/250f331c-acf3-4b28-b2a4-515a8cceb844-catalog-content\") pod \"redhat-operators-vsmhw\" (UID: \"250f331c-acf3-4b28-b2a4-515a8cceb844\") " pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.847178 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6vrt\" (UniqueName: \"kubernetes.io/projected/250f331c-acf3-4b28-b2a4-515a8cceb844-kube-api-access-m6vrt\") pod \"redhat-operators-vsmhw\" (UID: \"250f331c-acf3-4b28-b2a4-515a8cceb844\") " pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.891318 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:08 crc kubenswrapper[4646]: E1203 10:56:08.891678 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:09.391663845 +0000 UTC m=+145.854719980 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.947452 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.987807 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vlx92"] Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.989047 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:56:08 crc kubenswrapper[4646]: I1203 10:56:08.993411 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:08 crc kubenswrapper[4646]: E1203 10:56:08.993725 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:09.493715148 +0000 UTC m=+145.956771283 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.014199 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vlx92"] Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.095844 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.095974 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clt4h\" (UniqueName: \"kubernetes.io/projected/e253153a-bb9a-4324-86ba-6fd879a3fff9-kube-api-access-clt4h\") pod \"redhat-operators-vlx92\" (UID: \"e253153a-bb9a-4324-86ba-6fd879a3fff9\") " pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.096012 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e253153a-bb9a-4324-86ba-6fd879a3fff9-catalog-content\") pod \"redhat-operators-vlx92\" (UID: \"e253153a-bb9a-4324-86ba-6fd879a3fff9\") " pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.096069 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e253153a-bb9a-4324-86ba-6fd879a3fff9-utilities\") pod \"redhat-operators-vlx92\" (UID: \"e253153a-bb9a-4324-86ba-6fd879a3fff9\") " pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:56:09 crc kubenswrapper[4646]: E1203 10:56:09.096210 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:09.596195944 +0000 UTC m=+146.059252079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.172664 4646 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.184820 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ft6s8"] Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.197280 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e253153a-bb9a-4324-86ba-6fd879a3fff9-utilities\") pod \"redhat-operators-vlx92\" (UID: \"e253153a-bb9a-4324-86ba-6fd879a3fff9\") " pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.197427 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.197511 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clt4h\" (UniqueName: \"kubernetes.io/projected/e253153a-bb9a-4324-86ba-6fd879a3fff9-kube-api-access-clt4h\") pod \"redhat-operators-vlx92\" (UID: \"e253153a-bb9a-4324-86ba-6fd879a3fff9\") " pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.197554 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e253153a-bb9a-4324-86ba-6fd879a3fff9-catalog-content\") pod \"redhat-operators-vlx92\" (UID: \"e253153a-bb9a-4324-86ba-6fd879a3fff9\") " pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:56:09 crc kubenswrapper[4646]: E1203 10:56:09.198256 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:09.698236407 +0000 UTC m=+146.161292632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.198308 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e253153a-bb9a-4324-86ba-6fd879a3fff9-catalog-content\") pod \"redhat-operators-vlx92\" (UID: \"e253153a-bb9a-4324-86ba-6fd879a3fff9\") " pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.198502 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e253153a-bb9a-4324-86ba-6fd879a3fff9-utilities\") pod \"redhat-operators-vlx92\" (UID: \"e253153a-bb9a-4324-86ba-6fd879a3fff9\") " pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.229623 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clt4h\" (UniqueName: \"kubernetes.io/projected/e253153a-bb9a-4324-86ba-6fd879a3fff9-kube-api-access-clt4h\") pod \"redhat-operators-vlx92\" (UID: \"e253153a-bb9a-4324-86ba-6fd879a3fff9\") " pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.302053 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:09 crc kubenswrapper[4646]: E1203 10:56:09.302644 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:09.802627938 +0000 UTC m=+146.265684073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.319119 4646 patch_prober.go:28] interesting pod/console-operator-58897d9998-52ndk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.319168 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-52ndk" podUID="e6da1553-a91a-4055-af3d-bd30b20e6c0f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.17:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.319362 4646 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-9x4fj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.319428 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" podUID="cd573c9a-19ea-4975-9645-9d5b9e258d58" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.325510 4646 generic.go:334] "Generic (PLEG): container finished" podID="d1141f44-4444-47fb-878e-7a17db767b8d" containerID="f69d040041dfd1c522638e466cf336112bf5c7f43b79f49a4e6160f2ddf229df" exitCode=0 Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.325580 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkpc9" event={"ID":"d1141f44-4444-47fb-878e-7a17db767b8d","Type":"ContainerDied","Data":"f69d040041dfd1c522638e466cf336112bf5c7f43b79f49a4e6160f2ddf229df"} Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.327008 4646 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.327307 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.342843 4646 generic.go:334] "Generic (PLEG): container finished" podID="2b646c66-7457-42fe-abe5-0def6d0d3262" containerID="492ce31ba6978193f06ecdb65ff05603d5e24ea781aa7ac2d212f3902afba3f0" exitCode=0 Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.342929 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kv72s" event={"ID":"2b646c66-7457-42fe-abe5-0def6d0d3262","Type":"ContainerDied","Data":"492ce31ba6978193f06ecdb65ff05603d5e24ea781aa7ac2d212f3902afba3f0"} Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.376641 4646 generic.go:334] "Generic (PLEG): container finished" podID="b13d8714-e047-4380-b095-7f5aac84f021" containerID="31326048919f5e8ecb016a0b00ce9fc87aa78a5c06ac18e588412e93940a0249" exitCode=0 Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.376764 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p9lsq" event={"ID":"b13d8714-e047-4380-b095-7f5aac84f021","Type":"ContainerDied","Data":"31326048919f5e8ecb016a0b00ce9fc87aa78a5c06ac18e588412e93940a0249"} Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.412387 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:09 crc kubenswrapper[4646]: E1203 10:56:09.413118 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:09.913105082 +0000 UTC m=+146.376161217 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.444756 4646 generic.go:334] "Generic (PLEG): container finished" podID="d20d9814-aa42-4109-9740-d0e5e6205a00" containerID="d82a373f45467842e1c0037753d5e6c184d58e75507db49e79657d1bd0052da4" exitCode=0 Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.444884 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjl9g" event={"ID":"d20d9814-aa42-4109-9740-d0e5e6205a00","Type":"ContainerDied","Data":"d82a373f45467842e1c0037753d5e6c184d58e75507db49e79657d1bd0052da4"} Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.444912 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjl9g" event={"ID":"d20d9814-aa42-4109-9740-d0e5e6205a00","Type":"ContainerStarted","Data":"e9c7a8c9b8cf5a21833988a2e3ad4a8fcf3cae119cdf014cf47488f75b97e880"} Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.446082 4646 patch_prober.go:28] interesting pod/router-default-5444994796-7whtv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 10:56:09 crc kubenswrapper[4646]: [-]has-synced failed: reason withheld Dec 03 10:56:09 crc kubenswrapper[4646]: [+]process-running ok Dec 03 10:56:09 crc kubenswrapper[4646]: healthz check failed Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.446124 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7whtv" podUID="aa99705d-9e2f-4c32-85cb-04f0cde27411" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.474640 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.475829 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.476364 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vsmhw"] Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.486563 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.514944 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:09 crc kubenswrapper[4646]: E1203 10:56:09.515938 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:10.015917467 +0000 UTC m=+146.478973602 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.517326 4646 generic.go:334] "Generic (PLEG): container finished" podID="92bfb497-1e60-4973-a76c-0c686689cc53" containerID="4cefd6c44c3970872f8319f1ac85c862a6715a5836dd6cfb7c4eee8a6859a735" exitCode=0 Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.517447 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzw7n" event={"ID":"92bfb497-1e60-4973-a76c-0c686689cc53","Type":"ContainerDied","Data":"4cefd6c44c3970872f8319f1ac85c862a6715a5836dd6cfb7c4eee8a6859a735"} Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.528487 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.528547 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.546517 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-czsqx" event={"ID":"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6","Type":"ContainerStarted","Data":"a3ac1f5b67f78c601545b4a80ac0deac777e25be9c66cd5063efc9f7bf2b858f"} Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.564107 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ft6s8" event={"ID":"d4048178-b02e-472c-a4db-5520e198217c","Type":"ContainerStarted","Data":"a1e687a7af51e853e36f4259d0e389a02233f3351e45a99850f0f6e8b842f7cf"} Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.571636 4646 patch_prober.go:28] interesting pod/apiserver-76f77b778f-q7bzw container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 03 10:56:09 crc kubenswrapper[4646]: [+]log ok Dec 03 10:56:09 crc kubenswrapper[4646]: [+]etcd ok Dec 03 10:56:09 crc kubenswrapper[4646]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 03 10:56:09 crc kubenswrapper[4646]: [+]poststarthook/generic-apiserver-start-informers ok Dec 03 10:56:09 crc kubenswrapper[4646]: [+]poststarthook/max-in-flight-filter ok Dec 03 10:56:09 crc kubenswrapper[4646]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 03 10:56:09 crc kubenswrapper[4646]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 03 10:56:09 crc kubenswrapper[4646]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 03 10:56:09 crc kubenswrapper[4646]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Dec 03 10:56:09 crc kubenswrapper[4646]: [+]poststarthook/project.openshift.io-projectcache ok Dec 03 10:56:09 crc kubenswrapper[4646]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 03 10:56:09 crc kubenswrapper[4646]: [+]poststarthook/openshift.io-startinformers ok Dec 03 10:56:09 crc kubenswrapper[4646]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 03 10:56:09 crc kubenswrapper[4646]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 03 10:56:09 crc kubenswrapper[4646]: livez check failed Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.572127 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" podUID="9a0dc011-d040-4ec5-babf-d95b03ec699e" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.581043 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9x4fj" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.616261 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:09 crc kubenswrapper[4646]: E1203 10:56:09.617264 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:10.11725295 +0000 UTC m=+146.580309085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.718157 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:09 crc kubenswrapper[4646]: E1203 10:56:09.720444 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:10.220424026 +0000 UTC m=+146.683480161 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.755861 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-52ndk" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.823560 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.823623 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.823645 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.823679 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.823697 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:56:09 crc kubenswrapper[4646]: E1203 10:56:09.824617 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:10.32460143 +0000 UTC m=+146.787657555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.825569 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.839304 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.842016 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.844503 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.869720 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.912795 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-czsqx" podStartSLOduration=12.912775028 podStartE2EDuration="12.912775028s" podCreationTimestamp="2025-12-03 10:55:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:09.822738587 +0000 UTC m=+146.285794722" watchObservedRunningTime="2025-12-03 10:56:09.912775028 +0000 UTC m=+146.375831163" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.924705 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:09 crc kubenswrapper[4646]: E1203 10:56:09.925288 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 10:56:10.425250654 +0000 UTC m=+146.888306789 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.973043 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 10:56:09 crc kubenswrapper[4646]: I1203 10:56:09.986757 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.015152 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.016489 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.020552 4646 patch_prober.go:28] interesting pod/console-f9d7485db-rvbtr container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.020599 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-rvbtr" podUID="d109aa7c-89de-44b3-852c-b3c49b2bb2a0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.026248 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:10 crc kubenswrapper[4646]: E1203 10:56:10.026565 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 10:56:10.526552306 +0000 UTC m=+146.989608441 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k4mmg" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.084635 4646 patch_prober.go:28] interesting pod/downloads-7954f5f757-l2kck container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.084673 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-l2kck" podUID="54c0778a-7e52-4da6-ab86-6a3492eb278b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.084722 4646 patch_prober.go:28] interesting pod/downloads-7954f5f757-l2kck container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.084732 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-l2kck" podUID="54c0778a-7e52-4da6-ab86-6a3492eb278b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.088903 4646 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-03T10:56:09.172686108Z","Handler":null,"Name":""} Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.091482 4646 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.091503 4646 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.127241 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.174934 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.229511 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.250346 4646 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.250620 4646 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.256118 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vlx92"] Dec 03 10:56:10 crc kubenswrapper[4646]: W1203 10:56:10.274657 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode253153a_bb9a_4324_86ba_6fd879a3fff9.slice/crio-7f73e2f7e25a6fa23284eae31e7dd69b6be33cdd32093ee13ae1f954c08c975c WatchSource:0}: Error finding container 7f73e2f7e25a6fa23284eae31e7dd69b6be33cdd32093ee13ae1f954c08c975c: Status 404 returned error can't find the container with id 7f73e2f7e25a6fa23284eae31e7dd69b6be33cdd32093ee13ae1f954c08c975c Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.422704 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.426355 4646 patch_prober.go:28] interesting pod/router-default-5444994796-7whtv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 10:56:10 crc kubenswrapper[4646]: [-]has-synced failed: reason withheld Dec 03 10:56:10 crc kubenswrapper[4646]: [+]process-running ok Dec 03 10:56:10 crc kubenswrapper[4646]: healthz check failed Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.426403 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7whtv" podUID="aa99705d-9e2f-4c32-85cb-04f0cde27411" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.516842 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k4mmg\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.579795 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.593647 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.643548 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-czsqx" event={"ID":"dc66a521-1bd1-4e6a-8bc8-9cd1cd9a48b6","Type":"ContainerStarted","Data":"0f770aa09c4c4a6ab7f6768a72b9dc761dff0f76d37fa7ed9113d235874ba38c"} Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.652768 4646 generic.go:334] "Generic (PLEG): container finished" podID="d4048178-b02e-472c-a4db-5520e198217c" containerID="467ab9ea48488d24745018335241c53a98ef25a0ad8f6320ec1caf276b39d82f" exitCode=0 Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.652810 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ft6s8" event={"ID":"d4048178-b02e-472c-a4db-5520e198217c","Type":"ContainerDied","Data":"467ab9ea48488d24745018335241c53a98ef25a0ad8f6320ec1caf276b39d82f"} Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.662621 4646 generic.go:334] "Generic (PLEG): container finished" podID="250f331c-acf3-4b28-b2a4-515a8cceb844" containerID="a75b773fc6a2e0df7b0b6ab650b485a193bc7a61fa8320aae2b8c005e1c7dcc1" exitCode=0 Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.662704 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsmhw" event={"ID":"250f331c-acf3-4b28-b2a4-515a8cceb844","Type":"ContainerDied","Data":"a75b773fc6a2e0df7b0b6ab650b485a193bc7a61fa8320aae2b8c005e1c7dcc1"} Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.662738 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsmhw" event={"ID":"250f331c-acf3-4b28-b2a4-515a8cceb844","Type":"ContainerStarted","Data":"0eaf5f3c287678ceb54cf44f2f32c7ad804002b076b094183b514a348f31beeb"} Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.680909 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlx92" event={"ID":"e253153a-bb9a-4324-86ba-6fd879a3fff9","Type":"ContainerStarted","Data":"7f73e2f7e25a6fa23284eae31e7dd69b6be33cdd32093ee13ae1f954c08c975c"} Dec 03 10:56:10 crc kubenswrapper[4646]: I1203 10:56:10.713652 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j6mst" Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.112312 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k4mmg"] Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.425948 4646 patch_prober.go:28] interesting pod/router-default-5444994796-7whtv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 10:56:11 crc kubenswrapper[4646]: [-]has-synced failed: reason withheld Dec 03 10:56:11 crc kubenswrapper[4646]: [+]process-running ok Dec 03 10:56:11 crc kubenswrapper[4646]: healthz check failed Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.426236 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7whtv" podUID="aa99705d-9e2f-4c32-85cb-04f0cde27411" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.711419 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"f1e213a1c80bc66c914ef39a78dcf7c4cdf89d319fbc4b8ce69c233fed70b05d"} Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.711464 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e2d47c79efc2fa104be3501e8206f623a7070abe12f05f90ba3e771d57085ec3"} Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.769916 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"af011cb56a18eb53ca98abe59285f4bb1adb71fe0b19bd8c730b3ae7b3e99524"} Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.769994 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"15e22aac6614e172cff756bfc43b1b2fc74f60ff15785862877b962c3a77ff91"} Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.838668 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9d71d75f842db676c516f0eeb8d1e46dfad6e2102088cec1e9f80bd45266b581"} Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.838908 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c6e09eb2daf0459e0eda472a132295ee95873d66547956e1dd2a883471892e18"} Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.839639 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.901630 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.902285 4646 generic.go:334] "Generic (PLEG): container finished" podID="d8e77566-f3a2-4a10-84b5-4e6078bc6f23" containerID="011d2186daa42776238b6189b1866584dcfdc7148abe0731e3b9c1021271b63d" exitCode=0 Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.902356 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" event={"ID":"d8e77566-f3a2-4a10-84b5-4e6078bc6f23","Type":"ContainerDied","Data":"011d2186daa42776238b6189b1866584dcfdc7148abe0731e3b9c1021271b63d"} Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.950267 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" event={"ID":"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd","Type":"ContainerStarted","Data":"7b0242a4a54234730cd700a00a2ca7001496a1f04f110099d7cee9038608e060"} Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.950306 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" event={"ID":"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd","Type":"ContainerStarted","Data":"17c0251379134c085490348be935978ac89666a8e3a4371f516a4cae18e8eb4d"} Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.950563 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.960360 4646 generic.go:334] "Generic (PLEG): container finished" podID="e253153a-bb9a-4324-86ba-6fd879a3fff9" containerID="d28ef5e9644f869e00ceee30738fce53a9df58c5ec833668550e7b5964312962" exitCode=0 Dec 03 10:56:11 crc kubenswrapper[4646]: I1203 10:56:11.961472 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlx92" event={"ID":"e253153a-bb9a-4324-86ba-6fd879a3fff9","Type":"ContainerDied","Data":"d28ef5e9644f869e00ceee30738fce53a9df58c5ec833668550e7b5964312962"} Dec 03 10:56:12 crc kubenswrapper[4646]: I1203 10:56:12.043209 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" podStartSLOduration=130.043191371 podStartE2EDuration="2m10.043191371s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:11.989675174 +0000 UTC m=+148.452731299" watchObservedRunningTime="2025-12-03 10:56:12.043191371 +0000 UTC m=+148.506247506" Dec 03 10:56:12 crc kubenswrapper[4646]: I1203 10:56:12.437792 4646 patch_prober.go:28] interesting pod/router-default-5444994796-7whtv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 10:56:12 crc kubenswrapper[4646]: [-]has-synced failed: reason withheld Dec 03 10:56:12 crc kubenswrapper[4646]: [+]process-running ok Dec 03 10:56:12 crc kubenswrapper[4646]: healthz check failed Dec 03 10:56:12 crc kubenswrapper[4646]: I1203 10:56:12.437855 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7whtv" podUID="aa99705d-9e2f-4c32-85cb-04f0cde27411" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:12 crc kubenswrapper[4646]: I1203 10:56:12.769643 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 10:56:12 crc kubenswrapper[4646]: I1203 10:56:12.770318 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 10:56:12 crc kubenswrapper[4646]: I1203 10:56:12.777327 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 03 10:56:12 crc kubenswrapper[4646]: I1203 10:56:12.777871 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 03 10:56:12 crc kubenswrapper[4646]: I1203 10:56:12.793077 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 10:56:12 crc kubenswrapper[4646]: I1203 10:56:12.897227 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c53d8c48-57ef-4cef-b906-4ef9b226700d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c53d8c48-57ef-4cef-b906-4ef9b226700d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 10:56:12 crc kubenswrapper[4646]: I1203 10:56:12.897442 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c53d8c48-57ef-4cef-b906-4ef9b226700d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c53d8c48-57ef-4cef-b906-4ef9b226700d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 10:56:12 crc kubenswrapper[4646]: I1203 10:56:12.998631 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c53d8c48-57ef-4cef-b906-4ef9b226700d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c53d8c48-57ef-4cef-b906-4ef9b226700d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 10:56:12 crc kubenswrapper[4646]: I1203 10:56:12.998707 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c53d8c48-57ef-4cef-b906-4ef9b226700d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c53d8c48-57ef-4cef-b906-4ef9b226700d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 10:56:12 crc kubenswrapper[4646]: I1203 10:56:12.998836 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c53d8c48-57ef-4cef-b906-4ef9b226700d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c53d8c48-57ef-4cef-b906-4ef9b226700d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 10:56:13 crc kubenswrapper[4646]: I1203 10:56:13.021395 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c53d8c48-57ef-4cef-b906-4ef9b226700d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c53d8c48-57ef-4cef-b906-4ef9b226700d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 10:56:13 crc kubenswrapper[4646]: I1203 10:56:13.100166 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 10:56:13 crc kubenswrapper[4646]: I1203 10:56:13.423961 4646 patch_prober.go:28] interesting pod/router-default-5444994796-7whtv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 10:56:13 crc kubenswrapper[4646]: [-]has-synced failed: reason withheld Dec 03 10:56:13 crc kubenswrapper[4646]: [+]process-running ok Dec 03 10:56:13 crc kubenswrapper[4646]: healthz check failed Dec 03 10:56:13 crc kubenswrapper[4646]: I1203 10:56:13.424003 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7whtv" podUID="aa99705d-9e2f-4c32-85cb-04f0cde27411" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:13 crc kubenswrapper[4646]: I1203 10:56:13.452998 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" Dec 03 10:56:13 crc kubenswrapper[4646]: I1203 10:56:13.504104 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjwhf\" (UniqueName: \"kubernetes.io/projected/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-kube-api-access-tjwhf\") pod \"d8e77566-f3a2-4a10-84b5-4e6078bc6f23\" (UID: \"d8e77566-f3a2-4a10-84b5-4e6078bc6f23\") " Dec 03 10:56:13 crc kubenswrapper[4646]: I1203 10:56:13.504211 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-secret-volume\") pod \"d8e77566-f3a2-4a10-84b5-4e6078bc6f23\" (UID: \"d8e77566-f3a2-4a10-84b5-4e6078bc6f23\") " Dec 03 10:56:13 crc kubenswrapper[4646]: I1203 10:56:13.504255 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-config-volume\") pod \"d8e77566-f3a2-4a10-84b5-4e6078bc6f23\" (UID: \"d8e77566-f3a2-4a10-84b5-4e6078bc6f23\") " Dec 03 10:56:13 crc kubenswrapper[4646]: I1203 10:56:13.505679 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-config-volume" (OuterVolumeSpecName: "config-volume") pod "d8e77566-f3a2-4a10-84b5-4e6078bc6f23" (UID: "d8e77566-f3a2-4a10-84b5-4e6078bc6f23"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:56:13 crc kubenswrapper[4646]: I1203 10:56:13.510583 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-kube-api-access-tjwhf" (OuterVolumeSpecName: "kube-api-access-tjwhf") pod "d8e77566-f3a2-4a10-84b5-4e6078bc6f23" (UID: "d8e77566-f3a2-4a10-84b5-4e6078bc6f23"). InnerVolumeSpecName "kube-api-access-tjwhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:56:13 crc kubenswrapper[4646]: I1203 10:56:13.511053 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d8e77566-f3a2-4a10-84b5-4e6078bc6f23" (UID: "d8e77566-f3a2-4a10-84b5-4e6078bc6f23"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:56:13 crc kubenswrapper[4646]: I1203 10:56:13.606430 4646 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 10:56:13 crc kubenswrapper[4646]: I1203 10:56:13.606464 4646 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 10:56:13 crc kubenswrapper[4646]: I1203 10:56:13.606474 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjwhf\" (UniqueName: \"kubernetes.io/projected/d8e77566-f3a2-4a10-84b5-4e6078bc6f23-kube-api-access-tjwhf\") on node \"crc\" DevicePath \"\"" Dec 03 10:56:13 crc kubenswrapper[4646]: I1203 10:56:13.797854 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 10:56:14 crc kubenswrapper[4646]: I1203 10:56:14.031674 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" Dec 03 10:56:14 crc kubenswrapper[4646]: I1203 10:56:14.031722 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t" event={"ID":"d8e77566-f3a2-4a10-84b5-4e6078bc6f23","Type":"ContainerDied","Data":"d895cd81faca6bd4c4ff8eccaf84bf9257310e4b9166319dc9451e27ee4da175"} Dec 03 10:56:14 crc kubenswrapper[4646]: I1203 10:56:14.031773 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d895cd81faca6bd4c4ff8eccaf84bf9257310e4b9166319dc9451e27ee4da175" Dec 03 10:56:14 crc kubenswrapper[4646]: I1203 10:56:14.038627 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c53d8c48-57ef-4cef-b906-4ef9b226700d","Type":"ContainerStarted","Data":"197f159946b3348dbe9ee5adb539d6c34f054cdce253ae95dc11d5c79e1cbace"} Dec 03 10:56:14 crc kubenswrapper[4646]: I1203 10:56:14.375506 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 10:56:14 crc kubenswrapper[4646]: I1203 10:56:14.440689 4646 patch_prober.go:28] interesting pod/router-default-5444994796-7whtv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 10:56:14 crc kubenswrapper[4646]: [-]has-synced failed: reason withheld Dec 03 10:56:14 crc kubenswrapper[4646]: [+]process-running ok Dec 03 10:56:14 crc kubenswrapper[4646]: healthz check failed Dec 03 10:56:14 crc kubenswrapper[4646]: I1203 10:56:14.441021 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7whtv" podUID="aa99705d-9e2f-4c32-85cb-04f0cde27411" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:14 crc kubenswrapper[4646]: I1203 10:56:14.555622 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:56:14 crc kubenswrapper[4646]: I1203 10:56:14.573793 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-q7bzw" Dec 03 10:56:15 crc kubenswrapper[4646]: I1203 10:56:15.266258 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-hc78d" Dec 03 10:56:15 crc kubenswrapper[4646]: I1203 10:56:15.434052 4646 patch_prober.go:28] interesting pod/router-default-5444994796-7whtv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 10:56:15 crc kubenswrapper[4646]: [-]has-synced failed: reason withheld Dec 03 10:56:15 crc kubenswrapper[4646]: [+]process-running ok Dec 03 10:56:15 crc kubenswrapper[4646]: healthz check failed Dec 03 10:56:15 crc kubenswrapper[4646]: I1203 10:56:15.434108 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7whtv" podUID="aa99705d-9e2f-4c32-85cb-04f0cde27411" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:16 crc kubenswrapper[4646]: I1203 10:56:16.064578 4646 generic.go:334] "Generic (PLEG): container finished" podID="c53d8c48-57ef-4cef-b906-4ef9b226700d" containerID="306814cd460aae03e124117df9b5d1905b477f6426c12746ba56c423c36d2b4b" exitCode=0 Dec 03 10:56:16 crc kubenswrapper[4646]: I1203 10:56:16.064786 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c53d8c48-57ef-4cef-b906-4ef9b226700d","Type":"ContainerDied","Data":"306814cd460aae03e124117df9b5d1905b477f6426c12746ba56c423c36d2b4b"} Dec 03 10:56:16 crc kubenswrapper[4646]: I1203 10:56:16.425068 4646 patch_prober.go:28] interesting pod/router-default-5444994796-7whtv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 10:56:16 crc kubenswrapper[4646]: [-]has-synced failed: reason withheld Dec 03 10:56:16 crc kubenswrapper[4646]: [+]process-running ok Dec 03 10:56:16 crc kubenswrapper[4646]: healthz check failed Dec 03 10:56:16 crc kubenswrapper[4646]: I1203 10:56:16.425131 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7whtv" podUID="aa99705d-9e2f-4c32-85cb-04f0cde27411" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.178607 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 10:56:17 crc kubenswrapper[4646]: E1203 10:56:17.178849 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8e77566-f3a2-4a10-84b5-4e6078bc6f23" containerName="collect-profiles" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.178861 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8e77566-f3a2-4a10-84b5-4e6078bc6f23" containerName="collect-profiles" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.178960 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8e77566-f3a2-4a10-84b5-4e6078bc6f23" containerName="collect-profiles" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.179327 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.183272 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.184554 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.184684 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.282076 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0406784-811e-4bd7-872b-859ea611d1f6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d0406784-811e-4bd7-872b-859ea611d1f6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.282229 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d0406784-811e-4bd7-872b-859ea611d1f6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d0406784-811e-4bd7-872b-859ea611d1f6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.383926 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0406784-811e-4bd7-872b-859ea611d1f6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d0406784-811e-4bd7-872b-859ea611d1f6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.384257 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d0406784-811e-4bd7-872b-859ea611d1f6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d0406784-811e-4bd7-872b-859ea611d1f6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.384381 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d0406784-811e-4bd7-872b-859ea611d1f6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d0406784-811e-4bd7-872b-859ea611d1f6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.417037 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0406784-811e-4bd7-872b-859ea611d1f6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d0406784-811e-4bd7-872b-859ea611d1f6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.436998 4646 patch_prober.go:28] interesting pod/router-default-5444994796-7whtv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 10:56:17 crc kubenswrapper[4646]: [-]has-synced failed: reason withheld Dec 03 10:56:17 crc kubenswrapper[4646]: [+]process-running ok Dec 03 10:56:17 crc kubenswrapper[4646]: healthz check failed Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.437056 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7whtv" podUID="aa99705d-9e2f-4c32-85cb-04f0cde27411" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.502442 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.511828 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.588068 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c53d8c48-57ef-4cef-b906-4ef9b226700d-kubelet-dir\") pod \"c53d8c48-57ef-4cef-b906-4ef9b226700d\" (UID: \"c53d8c48-57ef-4cef-b906-4ef9b226700d\") " Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.588151 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c53d8c48-57ef-4cef-b906-4ef9b226700d-kube-api-access\") pod \"c53d8c48-57ef-4cef-b906-4ef9b226700d\" (UID: \"c53d8c48-57ef-4cef-b906-4ef9b226700d\") " Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.588169 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c53d8c48-57ef-4cef-b906-4ef9b226700d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c53d8c48-57ef-4cef-b906-4ef9b226700d" (UID: "c53d8c48-57ef-4cef-b906-4ef9b226700d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.589110 4646 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c53d8c48-57ef-4cef-b906-4ef9b226700d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.595732 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c53d8c48-57ef-4cef-b906-4ef9b226700d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c53d8c48-57ef-4cef-b906-4ef9b226700d" (UID: "c53d8c48-57ef-4cef-b906-4ef9b226700d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.690482 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c53d8c48-57ef-4cef-b906-4ef9b226700d-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 10:56:17 crc kubenswrapper[4646]: I1203 10:56:17.929900 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 10:56:18 crc kubenswrapper[4646]: I1203 10:56:18.127243 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 10:56:18 crc kubenswrapper[4646]: I1203 10:56:18.127246 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c53d8c48-57ef-4cef-b906-4ef9b226700d","Type":"ContainerDied","Data":"197f159946b3348dbe9ee5adb539d6c34f054cdce253ae95dc11d5c79e1cbace"} Dec 03 10:56:18 crc kubenswrapper[4646]: I1203 10:56:18.127293 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="197f159946b3348dbe9ee5adb539d6c34f054cdce253ae95dc11d5c79e1cbace" Dec 03 10:56:18 crc kubenswrapper[4646]: I1203 10:56:18.182024 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d0406784-811e-4bd7-872b-859ea611d1f6","Type":"ContainerStarted","Data":"160f1bfc21f1fc2ce7b32eedf18ca0cec48ead40eee4275ceeb6c3f316e74bcc"} Dec 03 10:56:18 crc kubenswrapper[4646]: I1203 10:56:18.427913 4646 patch_prober.go:28] interesting pod/router-default-5444994796-7whtv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 10:56:18 crc kubenswrapper[4646]: [-]has-synced failed: reason withheld Dec 03 10:56:18 crc kubenswrapper[4646]: [+]process-running ok Dec 03 10:56:18 crc kubenswrapper[4646]: healthz check failed Dec 03 10:56:18 crc kubenswrapper[4646]: I1203 10:56:18.427971 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7whtv" podUID="aa99705d-9e2f-4c32-85cb-04f0cde27411" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 10:56:19 crc kubenswrapper[4646]: I1203 10:56:19.197559 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d0406784-811e-4bd7-872b-859ea611d1f6","Type":"ContainerStarted","Data":"8967f0f3aadd0e80f7eb2442e4a5e6aa4f1957b08c1da5ab50b5b1450318eb5a"} Dec 03 10:56:19 crc kubenswrapper[4646]: I1203 10:56:19.425577 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:56:19 crc kubenswrapper[4646]: I1203 10:56:19.432445 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-7whtv" Dec 03 10:56:20 crc kubenswrapper[4646]: I1203 10:56:20.014709 4646 patch_prober.go:28] interesting pod/console-f9d7485db-rvbtr container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Dec 03 10:56:20 crc kubenswrapper[4646]: I1203 10:56:20.014795 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-rvbtr" podUID="d109aa7c-89de-44b3-852c-b3c49b2bb2a0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Dec 03 10:56:20 crc kubenswrapper[4646]: I1203 10:56:20.098607 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-l2kck" Dec 03 10:56:20 crc kubenswrapper[4646]: I1203 10:56:20.223023 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.223008508 podStartE2EDuration="3.223008508s" podCreationTimestamp="2025-12-03 10:56:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:20.220801325 +0000 UTC m=+156.683857460" watchObservedRunningTime="2025-12-03 10:56:20.223008508 +0000 UTC m=+156.686064643" Dec 03 10:56:21 crc kubenswrapper[4646]: I1203 10:56:21.246717 4646 generic.go:334] "Generic (PLEG): container finished" podID="d0406784-811e-4bd7-872b-859ea611d1f6" containerID="8967f0f3aadd0e80f7eb2442e4a5e6aa4f1957b08c1da5ab50b5b1450318eb5a" exitCode=0 Dec 03 10:56:21 crc kubenswrapper[4646]: I1203 10:56:21.246771 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d0406784-811e-4bd7-872b-859ea611d1f6","Type":"ContainerDied","Data":"8967f0f3aadd0e80f7eb2442e4a5e6aa4f1957b08c1da5ab50b5b1450318eb5a"} Dec 03 10:56:22 crc kubenswrapper[4646]: I1203 10:56:22.749481 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 10:56:22 crc kubenswrapper[4646]: I1203 10:56:22.763883 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0406784-811e-4bd7-872b-859ea611d1f6-kube-api-access\") pod \"d0406784-811e-4bd7-872b-859ea611d1f6\" (UID: \"d0406784-811e-4bd7-872b-859ea611d1f6\") " Dec 03 10:56:22 crc kubenswrapper[4646]: I1203 10:56:22.764024 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d0406784-811e-4bd7-872b-859ea611d1f6-kubelet-dir\") pod \"d0406784-811e-4bd7-872b-859ea611d1f6\" (UID: \"d0406784-811e-4bd7-872b-859ea611d1f6\") " Dec 03 10:56:22 crc kubenswrapper[4646]: I1203 10:56:22.764372 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0406784-811e-4bd7-872b-859ea611d1f6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d0406784-811e-4bd7-872b-859ea611d1f6" (UID: "d0406784-811e-4bd7-872b-859ea611d1f6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:56:22 crc kubenswrapper[4646]: I1203 10:56:22.776160 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0406784-811e-4bd7-872b-859ea611d1f6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d0406784-811e-4bd7-872b-859ea611d1f6" (UID: "d0406784-811e-4bd7-872b-859ea611d1f6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:56:22 crc kubenswrapper[4646]: I1203 10:56:22.866940 4646 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d0406784-811e-4bd7-872b-859ea611d1f6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 10:56:22 crc kubenswrapper[4646]: I1203 10:56:22.866975 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0406784-811e-4bd7-872b-859ea611d1f6-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 10:56:23 crc kubenswrapper[4646]: I1203 10:56:23.262104 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 10:56:23 crc kubenswrapper[4646]: I1203 10:56:23.262121 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d0406784-811e-4bd7-872b-859ea611d1f6","Type":"ContainerDied","Data":"160f1bfc21f1fc2ce7b32eedf18ca0cec48ead40eee4275ceeb6c3f316e74bcc"} Dec 03 10:56:23 crc kubenswrapper[4646]: I1203 10:56:23.262581 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="160f1bfc21f1fc2ce7b32eedf18ca0cec48ead40eee4275ceeb6c3f316e74bcc" Dec 03 10:56:24 crc kubenswrapper[4646]: I1203 10:56:24.486380 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs\") pod \"network-metrics-daemon-rs5rq\" (UID: \"51b171e4-cc29-4118-908c-d13825362e1c\") " pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:56:24 crc kubenswrapper[4646]: I1203 10:56:24.490553 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/51b171e4-cc29-4118-908c-d13825362e1c-metrics-certs\") pod \"network-metrics-daemon-rs5rq\" (UID: \"51b171e4-cc29-4118-908c-d13825362e1c\") " pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:56:24 crc kubenswrapper[4646]: I1203 10:56:24.576963 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs5rq" Dec 03 10:56:25 crc kubenswrapper[4646]: I1203 10:56:25.964274 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:56:25 crc kubenswrapper[4646]: I1203 10:56:25.964386 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:56:30 crc kubenswrapper[4646]: I1203 10:56:30.019909 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:56:30 crc kubenswrapper[4646]: I1203 10:56:30.025672 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 10:56:30 crc kubenswrapper[4646]: I1203 10:56:30.600650 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 10:56:40 crc kubenswrapper[4646]: I1203 10:56:40.491917 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2vh6v" Dec 03 10:56:45 crc kubenswrapper[4646]: E1203 10:56:45.491553 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 10:56:45 crc kubenswrapper[4646]: E1203 10:56:45.492572 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-crlwl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ft6s8_openshift-marketplace(d4048178-b02e-472c-a4db-5520e198217c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 10:56:45 crc kubenswrapper[4646]: E1203 10:56:45.493913 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-ft6s8" podUID="d4048178-b02e-472c-a4db-5520e198217c" Dec 03 10:56:45 crc kubenswrapper[4646]: I1203 10:56:45.935166 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-rs5rq"] Dec 03 10:56:45 crc kubenswrapper[4646]: W1203 10:56:45.941700 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51b171e4_cc29_4118_908c_d13825362e1c.slice/crio-d90afca4a169c4e66a25509dd11f5f2511714754e5838c19b21aa546fc018e35 WatchSource:0}: Error finding container d90afca4a169c4e66a25509dd11f5f2511714754e5838c19b21aa546fc018e35: Status 404 returned error can't find the container with id d90afca4a169c4e66a25509dd11f5f2511714754e5838c19b21aa546fc018e35 Dec 03 10:56:46 crc kubenswrapper[4646]: I1203 10:56:46.399193 4646 generic.go:334] "Generic (PLEG): container finished" podID="d1141f44-4444-47fb-878e-7a17db767b8d" containerID="7f97b5c971fc00c9628d805dd112cffaee6de0097a01311b4b0cdb7f949b5085" exitCode=0 Dec 03 10:56:46 crc kubenswrapper[4646]: I1203 10:56:46.399289 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkpc9" event={"ID":"d1141f44-4444-47fb-878e-7a17db767b8d","Type":"ContainerDied","Data":"7f97b5c971fc00c9628d805dd112cffaee6de0097a01311b4b0cdb7f949b5085"} Dec 03 10:56:46 crc kubenswrapper[4646]: I1203 10:56:46.402006 4646 generic.go:334] "Generic (PLEG): container finished" podID="2b646c66-7457-42fe-abe5-0def6d0d3262" containerID="3da029f6f0eeaff30acf62ee29ad9da10abc22622f09927720d03d14b88cf5dc" exitCode=0 Dec 03 10:56:46 crc kubenswrapper[4646]: I1203 10:56:46.402433 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kv72s" event={"ID":"2b646c66-7457-42fe-abe5-0def6d0d3262","Type":"ContainerDied","Data":"3da029f6f0eeaff30acf62ee29ad9da10abc22622f09927720d03d14b88cf5dc"} Dec 03 10:56:46 crc kubenswrapper[4646]: I1203 10:56:46.406016 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsmhw" event={"ID":"250f331c-acf3-4b28-b2a4-515a8cceb844","Type":"ContainerStarted","Data":"b48c86e7cc849df6c509ba02fd6c836621c1d2cf0e125e46861115af2795ad36"} Dec 03 10:56:46 crc kubenswrapper[4646]: I1203 10:56:46.408150 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlx92" event={"ID":"e253153a-bb9a-4324-86ba-6fd879a3fff9","Type":"ContainerStarted","Data":"a53a3e6f143876296e2cb5b04605a8f873b8e93540fd34ef91c7662bb747b9ec"} Dec 03 10:56:46 crc kubenswrapper[4646]: I1203 10:56:46.411083 4646 generic.go:334] "Generic (PLEG): container finished" podID="b13d8714-e047-4380-b095-7f5aac84f021" containerID="d83e90241c43fa5c9ec4c4dc28ad273aadf58f89b0ad1edcc0909d092a56b2a2" exitCode=0 Dec 03 10:56:46 crc kubenswrapper[4646]: I1203 10:56:46.411144 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p9lsq" event={"ID":"b13d8714-e047-4380-b095-7f5aac84f021","Type":"ContainerDied","Data":"d83e90241c43fa5c9ec4c4dc28ad273aadf58f89b0ad1edcc0909d092a56b2a2"} Dec 03 10:56:46 crc kubenswrapper[4646]: I1203 10:56:46.422888 4646 generic.go:334] "Generic (PLEG): container finished" podID="d20d9814-aa42-4109-9740-d0e5e6205a00" containerID="14b48e0f62154c7f2c9b9a3758864283bee56cea044cc60ede045e4395aa6cdb" exitCode=0 Dec 03 10:56:46 crc kubenswrapper[4646]: I1203 10:56:46.423017 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjl9g" event={"ID":"d20d9814-aa42-4109-9740-d0e5e6205a00","Type":"ContainerDied","Data":"14b48e0f62154c7f2c9b9a3758864283bee56cea044cc60ede045e4395aa6cdb"} Dec 03 10:56:46 crc kubenswrapper[4646]: I1203 10:56:46.425423 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzw7n" event={"ID":"92bfb497-1e60-4973-a76c-0c686689cc53","Type":"ContainerStarted","Data":"a4af1c9e511b645b729218dee2207679175801e838516dfc95116699413940ef"} Dec 03 10:56:46 crc kubenswrapper[4646]: I1203 10:56:46.426892 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" event={"ID":"51b171e4-cc29-4118-908c-d13825362e1c","Type":"ContainerStarted","Data":"d90afca4a169c4e66a25509dd11f5f2511714754e5838c19b21aa546fc018e35"} Dec 03 10:56:46 crc kubenswrapper[4646]: E1203 10:56:46.428511 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ft6s8" podUID="d4048178-b02e-472c-a4db-5520e198217c" Dec 03 10:56:47 crc kubenswrapper[4646]: I1203 10:56:47.432059 4646 generic.go:334] "Generic (PLEG): container finished" podID="92bfb497-1e60-4973-a76c-0c686689cc53" containerID="a4af1c9e511b645b729218dee2207679175801e838516dfc95116699413940ef" exitCode=0 Dec 03 10:56:47 crc kubenswrapper[4646]: I1203 10:56:47.432143 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzw7n" event={"ID":"92bfb497-1e60-4973-a76c-0c686689cc53","Type":"ContainerDied","Data":"a4af1c9e511b645b729218dee2207679175801e838516dfc95116699413940ef"} Dec 03 10:56:47 crc kubenswrapper[4646]: I1203 10:56:47.433458 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" event={"ID":"51b171e4-cc29-4118-908c-d13825362e1c","Type":"ContainerStarted","Data":"4d4b35ffb27996f833238bf09e67c05c14e98e5ca82d802871d265f6bcdf5e1a"} Dec 03 10:56:47 crc kubenswrapper[4646]: I1203 10:56:47.447810 4646 generic.go:334] "Generic (PLEG): container finished" podID="250f331c-acf3-4b28-b2a4-515a8cceb844" containerID="b48c86e7cc849df6c509ba02fd6c836621c1d2cf0e125e46861115af2795ad36" exitCode=0 Dec 03 10:56:47 crc kubenswrapper[4646]: I1203 10:56:47.449108 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsmhw" event={"ID":"250f331c-acf3-4b28-b2a4-515a8cceb844","Type":"ContainerDied","Data":"b48c86e7cc849df6c509ba02fd6c836621c1d2cf0e125e46861115af2795ad36"} Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.437748 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nb6wv"] Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.454507 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-rs5rq" event={"ID":"51b171e4-cc29-4118-908c-d13825362e1c","Type":"ContainerStarted","Data":"77a927bbf08948c00741f57be5c5565797471876e1b5fc8763ae869893e34df3"} Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.457781 4646 generic.go:334] "Generic (PLEG): container finished" podID="e253153a-bb9a-4324-86ba-6fd879a3fff9" containerID="a53a3e6f143876296e2cb5b04605a8f873b8e93540fd34ef91c7662bb747b9ec" exitCode=0 Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.457829 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlx92" event={"ID":"e253153a-bb9a-4324-86ba-6fd879a3fff9","Type":"ContainerDied","Data":"a53a3e6f143876296e2cb5b04605a8f873b8e93540fd34ef91c7662bb747b9ec"} Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.486777 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-rs5rq" podStartSLOduration=166.486755376 podStartE2EDuration="2m46.486755376s" podCreationTimestamp="2025-12-03 10:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:56:48.482750822 +0000 UTC m=+184.945806957" watchObservedRunningTime="2025-12-03 10:56:48.486755376 +0000 UTC m=+184.949811511" Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.597475 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 10:56:48 crc kubenswrapper[4646]: E1203 10:56:48.597708 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0406784-811e-4bd7-872b-859ea611d1f6" containerName="pruner" Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.597727 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0406784-811e-4bd7-872b-859ea611d1f6" containerName="pruner" Dec 03 10:56:48 crc kubenswrapper[4646]: E1203 10:56:48.597754 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c53d8c48-57ef-4cef-b906-4ef9b226700d" containerName="pruner" Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.597762 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="c53d8c48-57ef-4cef-b906-4ef9b226700d" containerName="pruner" Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.597870 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0406784-811e-4bd7-872b-859ea611d1f6" containerName="pruner" Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.597886 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="c53d8c48-57ef-4cef-b906-4ef9b226700d" containerName="pruner" Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.598232 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.605037 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.605450 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.616031 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/702787d9-af42-442e-83eb-9b05c8b7caf8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"702787d9-af42-442e-83eb-9b05c8b7caf8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.616079 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/702787d9-af42-442e-83eb-9b05c8b7caf8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"702787d9-af42-442e-83eb-9b05c8b7caf8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.627326 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.716957 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/702787d9-af42-442e-83eb-9b05c8b7caf8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"702787d9-af42-442e-83eb-9b05c8b7caf8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.717005 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/702787d9-af42-442e-83eb-9b05c8b7caf8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"702787d9-af42-442e-83eb-9b05c8b7caf8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.717134 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/702787d9-af42-442e-83eb-9b05c8b7caf8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"702787d9-af42-442e-83eb-9b05c8b7caf8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.744049 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/702787d9-af42-442e-83eb-9b05c8b7caf8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"702787d9-af42-442e-83eb-9b05c8b7caf8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 10:56:48 crc kubenswrapper[4646]: I1203 10:56:48.915590 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 10:56:50 crc kubenswrapper[4646]: I1203 10:56:50.298070 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 10:56:51 crc kubenswrapper[4646]: I1203 10:56:51.399917 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 10:56:51 crc kubenswrapper[4646]: I1203 10:56:51.474281 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"702787d9-af42-442e-83eb-9b05c8b7caf8","Type":"ContainerStarted","Data":"1a99050db1978a06b23e31556640944cd5f91a829842230aafbd323e546be8b7"} Dec 03 10:56:52 crc kubenswrapper[4646]: I1203 10:56:52.489015 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p9lsq" event={"ID":"b13d8714-e047-4380-b095-7f5aac84f021","Type":"ContainerStarted","Data":"7d94c7c40f614008d49a7ee595dd5336eaafd156f1fc0cf9b92e0c291d1103f0"} Dec 03 10:56:53 crc kubenswrapper[4646]: I1203 10:56:53.511271 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p9lsq" podStartSLOduration=6.971575128 podStartE2EDuration="48.511241207s" podCreationTimestamp="2025-12-03 10:56:05 +0000 UTC" firstStartedPulling="2025-12-03 10:56:09.38150735 +0000 UTC m=+145.844563485" lastFinishedPulling="2025-12-03 10:56:50.921173429 +0000 UTC m=+187.384229564" observedRunningTime="2025-12-03 10:56:53.510136276 +0000 UTC m=+189.973192421" watchObservedRunningTime="2025-12-03 10:56:53.511241207 +0000 UTC m=+189.974297342" Dec 03 10:56:53 crc kubenswrapper[4646]: I1203 10:56:53.574783 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 10:56:53 crc kubenswrapper[4646]: I1203 10:56:53.575736 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 10:56:53 crc kubenswrapper[4646]: I1203 10:56:53.585399 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 10:56:53 crc kubenswrapper[4646]: I1203 10:56:53.681652 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9030be92-e97b-4760-a4cf-abe87c241ec6-var-lock\") pod \"installer-9-crc\" (UID: \"9030be92-e97b-4760-a4cf-abe87c241ec6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 10:56:53 crc kubenswrapper[4646]: I1203 10:56:53.681717 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9030be92-e97b-4760-a4cf-abe87c241ec6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"9030be92-e97b-4760-a4cf-abe87c241ec6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 10:56:53 crc kubenswrapper[4646]: I1203 10:56:53.681754 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9030be92-e97b-4760-a4cf-abe87c241ec6-kube-api-access\") pod \"installer-9-crc\" (UID: \"9030be92-e97b-4760-a4cf-abe87c241ec6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 10:56:53 crc kubenswrapper[4646]: I1203 10:56:53.782907 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9030be92-e97b-4760-a4cf-abe87c241ec6-kube-api-access\") pod \"installer-9-crc\" (UID: \"9030be92-e97b-4760-a4cf-abe87c241ec6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 10:56:53 crc kubenswrapper[4646]: I1203 10:56:53.783002 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9030be92-e97b-4760-a4cf-abe87c241ec6-var-lock\") pod \"installer-9-crc\" (UID: \"9030be92-e97b-4760-a4cf-abe87c241ec6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 10:56:53 crc kubenswrapper[4646]: I1203 10:56:53.783050 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9030be92-e97b-4760-a4cf-abe87c241ec6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"9030be92-e97b-4760-a4cf-abe87c241ec6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 10:56:53 crc kubenswrapper[4646]: I1203 10:56:53.783118 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9030be92-e97b-4760-a4cf-abe87c241ec6-var-lock\") pod \"installer-9-crc\" (UID: \"9030be92-e97b-4760-a4cf-abe87c241ec6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 10:56:53 crc kubenswrapper[4646]: I1203 10:56:53.783136 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9030be92-e97b-4760-a4cf-abe87c241ec6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"9030be92-e97b-4760-a4cf-abe87c241ec6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 10:56:53 crc kubenswrapper[4646]: I1203 10:56:53.802824 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9030be92-e97b-4760-a4cf-abe87c241ec6-kube-api-access\") pod \"installer-9-crc\" (UID: \"9030be92-e97b-4760-a4cf-abe87c241ec6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 10:56:53 crc kubenswrapper[4646]: I1203 10:56:53.903170 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 10:56:55 crc kubenswrapper[4646]: I1203 10:56:55.660631 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:56:55 crc kubenswrapper[4646]: I1203 10:56:55.660943 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:56:55 crc kubenswrapper[4646]: I1203 10:56:55.714655 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:56:55 crc kubenswrapper[4646]: I1203 10:56:55.964537 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:56:55 crc kubenswrapper[4646]: I1203 10:56:55.964588 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.270152 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 10:57:02 crc kubenswrapper[4646]: W1203 10:57:02.276648 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod9030be92_e97b_4760_a4cf_abe87c241ec6.slice/crio-3fce1dd1b54f4c11cf63f283ea671167c74d39e2f4821c842396ecf3afb00f19 WatchSource:0}: Error finding container 3fce1dd1b54f4c11cf63f283ea671167c74d39e2f4821c842396ecf3afb00f19: Status 404 returned error can't find the container with id 3fce1dd1b54f4c11cf63f283ea671167c74d39e2f4821c842396ecf3afb00f19 Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.545277 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kv72s" event={"ID":"2b646c66-7457-42fe-abe5-0def6d0d3262","Type":"ContainerStarted","Data":"bbfc33fc24f85c44898bf009cb81a717db6413fdb2ff6f937c1b0ef35b54fed0"} Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.547496 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsmhw" event={"ID":"250f331c-acf3-4b28-b2a4-515a8cceb844","Type":"ContainerStarted","Data":"d3af0c2b2856937cd01de17968d8a6e1f48a3e514fdd28700de3b42ce3aa34df"} Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.549368 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlx92" event={"ID":"e253153a-bb9a-4324-86ba-6fd879a3fff9","Type":"ContainerStarted","Data":"18980927df10a0f42dcdff7fd9355ed665533165f0646eff05a8e4e75217ab18"} Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.551382 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzw7n" event={"ID":"92bfb497-1e60-4973-a76c-0c686689cc53","Type":"ContainerStarted","Data":"25b4229b499c3a463ea2f7438e29c2e548681dc050d343482507ff659f409d26"} Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.553149 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"702787d9-af42-442e-83eb-9b05c8b7caf8","Type":"ContainerStarted","Data":"9c2c5716f74eab1a4923d7ada51237a720f57421042db99bbb43394d5cfb9a8c"} Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.554023 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9030be92-e97b-4760-a4cf-abe87c241ec6","Type":"ContainerStarted","Data":"3fce1dd1b54f4c11cf63f283ea671167c74d39e2f4821c842396ecf3afb00f19"} Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.555716 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkpc9" event={"ID":"d1141f44-4444-47fb-878e-7a17db767b8d","Type":"ContainerStarted","Data":"659b6c320ce5f4246e99503f13f50a49dcf9ed082c911dd7b9d433cff30b3fd0"} Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.557832 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjl9g" event={"ID":"d20d9814-aa42-4109-9740-d0e5e6205a00","Type":"ContainerStarted","Data":"499dae3d8c1f749985ebe71582fe19a396f8a13081554ce6a9fec1916437e37f"} Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.559730 4646 generic.go:334] "Generic (PLEG): container finished" podID="d4048178-b02e-472c-a4db-5520e198217c" containerID="8079cf5ed9c0c19029e29c8582d726e72791785d01406d469278505e9f35daae" exitCode=0 Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.559767 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ft6s8" event={"ID":"d4048178-b02e-472c-a4db-5520e198217c","Type":"ContainerDied","Data":"8079cf5ed9c0c19029e29c8582d726e72791785d01406d469278505e9f35daae"} Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.569617 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kv72s" podStartSLOduration=5.037121718 podStartE2EDuration="57.569604454s" podCreationTimestamp="2025-12-03 10:56:05 +0000 UTC" firstStartedPulling="2025-12-03 10:56:09.359733148 +0000 UTC m=+145.822789283" lastFinishedPulling="2025-12-03 10:57:01.892215884 +0000 UTC m=+198.355272019" observedRunningTime="2025-12-03 10:57:02.568157952 +0000 UTC m=+199.031214087" watchObservedRunningTime="2025-12-03 10:57:02.569604454 +0000 UTC m=+199.032660589" Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.617115 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rzw7n" podStartSLOduration=7.016543082 podStartE2EDuration="57.617098843s" podCreationTimestamp="2025-12-03 10:56:05 +0000 UTC" firstStartedPulling="2025-12-03 10:56:09.55067521 +0000 UTC m=+146.013731335" lastFinishedPulling="2025-12-03 10:57:00.151230941 +0000 UTC m=+196.614287096" observedRunningTime="2025-12-03 10:57:02.596203561 +0000 UTC m=+199.059259696" watchObservedRunningTime="2025-12-03 10:57:02.617098843 +0000 UTC m=+199.080154978" Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.617585 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vsmhw" podStartSLOduration=4.795672226 podStartE2EDuration="54.617581257s" podCreationTimestamp="2025-12-03 10:56:08 +0000 UTC" firstStartedPulling="2025-12-03 10:56:10.666683712 +0000 UTC m=+147.129739847" lastFinishedPulling="2025-12-03 10:57:00.488592743 +0000 UTC m=+196.951648878" observedRunningTime="2025-12-03 10:57:02.616426174 +0000 UTC m=+199.079482309" watchObservedRunningTime="2025-12-03 10:57:02.617581257 +0000 UTC m=+199.080637392" Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.640281 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vkpc9" podStartSLOduration=5.98820363 podStartE2EDuration="57.640258941s" podCreationTimestamp="2025-12-03 10:56:05 +0000 UTC" firstStartedPulling="2025-12-03 10:56:09.326674364 +0000 UTC m=+145.789730499" lastFinishedPulling="2025-12-03 10:57:00.978729675 +0000 UTC m=+197.441785810" observedRunningTime="2025-12-03 10:57:02.640255581 +0000 UTC m=+199.103311716" watchObservedRunningTime="2025-12-03 10:57:02.640258941 +0000 UTC m=+199.103315076" Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.659772 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=14.659756333 podStartE2EDuration="14.659756333s" podCreationTimestamp="2025-12-03 10:56:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:57:02.657038515 +0000 UTC m=+199.120094650" watchObservedRunningTime="2025-12-03 10:57:02.659756333 +0000 UTC m=+199.122812468" Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.707219 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xjl9g" podStartSLOduration=11.465301472 podStartE2EDuration="55.707202861s" podCreationTimestamp="2025-12-03 10:56:07 +0000 UTC" firstStartedPulling="2025-12-03 10:56:09.473780884 +0000 UTC m=+145.936837019" lastFinishedPulling="2025-12-03 10:56:53.715682273 +0000 UTC m=+190.178738408" observedRunningTime="2025-12-03 10:57:02.678504074 +0000 UTC m=+199.141560209" watchObservedRunningTime="2025-12-03 10:57:02.707202861 +0000 UTC m=+199.170258996" Dec 03 10:57:02 crc kubenswrapper[4646]: I1203 10:57:02.731643 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vlx92" podStartSLOduration=4.80764942 podStartE2EDuration="54.731624185s" podCreationTimestamp="2025-12-03 10:56:08 +0000 UTC" firstStartedPulling="2025-12-03 10:56:11.976139317 +0000 UTC m=+148.439195452" lastFinishedPulling="2025-12-03 10:57:01.900114072 +0000 UTC m=+198.363170217" observedRunningTime="2025-12-03 10:57:02.73143173 +0000 UTC m=+199.194487865" watchObservedRunningTime="2025-12-03 10:57:02.731624185 +0000 UTC m=+199.194680320" Dec 03 10:57:03 crc kubenswrapper[4646]: I1203 10:57:03.585948 4646 generic.go:334] "Generic (PLEG): container finished" podID="702787d9-af42-442e-83eb-9b05c8b7caf8" containerID="9c2c5716f74eab1a4923d7ada51237a720f57421042db99bbb43394d5cfb9a8c" exitCode=0 Dec 03 10:57:03 crc kubenswrapper[4646]: I1203 10:57:03.586177 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"702787d9-af42-442e-83eb-9b05c8b7caf8","Type":"ContainerDied","Data":"9c2c5716f74eab1a4923d7ada51237a720f57421042db99bbb43394d5cfb9a8c"} Dec 03 10:57:03 crc kubenswrapper[4646]: I1203 10:57:03.589666 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ft6s8" event={"ID":"d4048178-b02e-472c-a4db-5520e198217c","Type":"ContainerStarted","Data":"8ba10de4e74a0170bb5411d0efa011e0723e1ed26d850f475e9dc95c93f8562e"} Dec 03 10:57:03 crc kubenswrapper[4646]: I1203 10:57:03.591013 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9030be92-e97b-4760-a4cf-abe87c241ec6","Type":"ContainerStarted","Data":"2c5fbb57f828b7e822c0fbf9cce6aa353afa23a324bdad1c6b451d0de880d21a"} Dec 03 10:57:03 crc kubenswrapper[4646]: I1203 10:57:03.662591 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ft6s8" podStartSLOduration=4.404847447 podStartE2EDuration="56.662572145s" podCreationTimestamp="2025-12-03 10:56:07 +0000 UTC" firstStartedPulling="2025-12-03 10:56:10.655580975 +0000 UTC m=+147.118637110" lastFinishedPulling="2025-12-03 10:57:02.913305673 +0000 UTC m=+199.376361808" observedRunningTime="2025-12-03 10:57:03.642885807 +0000 UTC m=+200.105941942" watchObservedRunningTime="2025-12-03 10:57:03.662572145 +0000 UTC m=+200.125628280" Dec 03 10:57:04 crc kubenswrapper[4646]: I1203 10:57:04.902604 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 10:57:04 crc kubenswrapper[4646]: I1203 10:57:04.919217 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=11.919201583 podStartE2EDuration="11.919201583s" podCreationTimestamp="2025-12-03 10:56:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:57:03.664462179 +0000 UTC m=+200.127518314" watchObservedRunningTime="2025-12-03 10:57:04.919201583 +0000 UTC m=+201.382257718" Dec 03 10:57:04 crc kubenswrapper[4646]: I1203 10:57:04.937662 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/702787d9-af42-442e-83eb-9b05c8b7caf8-kube-api-access\") pod \"702787d9-af42-442e-83eb-9b05c8b7caf8\" (UID: \"702787d9-af42-442e-83eb-9b05c8b7caf8\") " Dec 03 10:57:04 crc kubenswrapper[4646]: I1203 10:57:04.937722 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/702787d9-af42-442e-83eb-9b05c8b7caf8-kubelet-dir\") pod \"702787d9-af42-442e-83eb-9b05c8b7caf8\" (UID: \"702787d9-af42-442e-83eb-9b05c8b7caf8\") " Dec 03 10:57:04 crc kubenswrapper[4646]: I1203 10:57:04.937847 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/702787d9-af42-442e-83eb-9b05c8b7caf8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "702787d9-af42-442e-83eb-9b05c8b7caf8" (UID: "702787d9-af42-442e-83eb-9b05c8b7caf8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:57:04 crc kubenswrapper[4646]: I1203 10:57:04.938043 4646 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/702787d9-af42-442e-83eb-9b05c8b7caf8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:04 crc kubenswrapper[4646]: I1203 10:57:04.944461 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/702787d9-af42-442e-83eb-9b05c8b7caf8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "702787d9-af42-442e-83eb-9b05c8b7caf8" (UID: "702787d9-af42-442e-83eb-9b05c8b7caf8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:57:05 crc kubenswrapper[4646]: I1203 10:57:05.039219 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/702787d9-af42-442e-83eb-9b05c8b7caf8-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:05 crc kubenswrapper[4646]: I1203 10:57:05.602492 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"702787d9-af42-442e-83eb-9b05c8b7caf8","Type":"ContainerDied","Data":"1a99050db1978a06b23e31556640944cd5f91a829842230aafbd323e546be8b7"} Dec 03 10:57:05 crc kubenswrapper[4646]: I1203 10:57:05.602529 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 10:57:05 crc kubenswrapper[4646]: I1203 10:57:05.602541 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a99050db1978a06b23e31556640944cd5f91a829842230aafbd323e546be8b7" Dec 03 10:57:05 crc kubenswrapper[4646]: I1203 10:57:05.705875 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:57:05 crc kubenswrapper[4646]: I1203 10:57:05.755323 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:57:05 crc kubenswrapper[4646]: I1203 10:57:05.755388 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:57:05 crc kubenswrapper[4646]: I1203 10:57:05.797092 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:57:05 crc kubenswrapper[4646]: I1203 10:57:05.929329 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:57:05 crc kubenswrapper[4646]: I1203 10:57:05.929415 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:57:05 crc kubenswrapper[4646]: I1203 10:57:05.966807 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:57:06 crc kubenswrapper[4646]: I1203 10:57:06.243511 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:57:06 crc kubenswrapper[4646]: I1203 10:57:06.243567 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:57:06 crc kubenswrapper[4646]: I1203 10:57:06.282740 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:57:07 crc kubenswrapper[4646]: I1203 10:57:07.770106 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:57:07 crc kubenswrapper[4646]: I1203 10:57:07.770156 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:57:07 crc kubenswrapper[4646]: I1203 10:57:07.811986 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:57:08 crc kubenswrapper[4646]: I1203 10:57:08.264196 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:57:08 crc kubenswrapper[4646]: I1203 10:57:08.264283 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:57:08 crc kubenswrapper[4646]: I1203 10:57:08.310432 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:57:08 crc kubenswrapper[4646]: I1203 10:57:08.661857 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:57:08 crc kubenswrapper[4646]: I1203 10:57:08.689271 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:57:08 crc kubenswrapper[4646]: I1203 10:57:08.948548 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:57:08 crc kubenswrapper[4646]: I1203 10:57:08.949562 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:57:09 crc kubenswrapper[4646]: I1203 10:57:09.011218 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:57:09 crc kubenswrapper[4646]: I1203 10:57:09.327722 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:57:09 crc kubenswrapper[4646]: I1203 10:57:09.327778 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:57:09 crc kubenswrapper[4646]: I1203 10:57:09.373939 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:57:09 crc kubenswrapper[4646]: I1203 10:57:09.674270 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:57:09 crc kubenswrapper[4646]: I1203 10:57:09.696981 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:57:11 crc kubenswrapper[4646]: I1203 10:57:11.389886 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ft6s8"] Dec 03 10:57:11 crc kubenswrapper[4646]: I1203 10:57:11.390395 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ft6s8" podUID="d4048178-b02e-472c-a4db-5520e198217c" containerName="registry-server" containerID="cri-o://8ba10de4e74a0170bb5411d0efa011e0723e1ed26d850f475e9dc95c93f8562e" gracePeriod=2 Dec 03 10:57:12 crc kubenswrapper[4646]: I1203 10:57:12.396763 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vlx92"] Dec 03 10:57:12 crc kubenswrapper[4646]: I1203 10:57:12.397091 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vlx92" podUID="e253153a-bb9a-4324-86ba-6fd879a3fff9" containerName="registry-server" containerID="cri-o://18980927df10a0f42dcdff7fd9355ed665533165f0646eff05a8e4e75217ab18" gracePeriod=2 Dec 03 10:57:12 crc kubenswrapper[4646]: I1203 10:57:12.651271 4646 generic.go:334] "Generic (PLEG): container finished" podID="d4048178-b02e-472c-a4db-5520e198217c" containerID="8ba10de4e74a0170bb5411d0efa011e0723e1ed26d850f475e9dc95c93f8562e" exitCode=0 Dec 03 10:57:12 crc kubenswrapper[4646]: I1203 10:57:12.651693 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ft6s8" event={"ID":"d4048178-b02e-472c-a4db-5520e198217c","Type":"ContainerDied","Data":"8ba10de4e74a0170bb5411d0efa011e0723e1ed26d850f475e9dc95c93f8562e"} Dec 03 10:57:12 crc kubenswrapper[4646]: I1203 10:57:12.655007 4646 generic.go:334] "Generic (PLEG): container finished" podID="e253153a-bb9a-4324-86ba-6fd879a3fff9" containerID="18980927df10a0f42dcdff7fd9355ed665533165f0646eff05a8e4e75217ab18" exitCode=0 Dec 03 10:57:12 crc kubenswrapper[4646]: I1203 10:57:12.655033 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlx92" event={"ID":"e253153a-bb9a-4324-86ba-6fd879a3fff9","Type":"ContainerDied","Data":"18980927df10a0f42dcdff7fd9355ed665533165f0646eff05a8e4e75217ab18"} Dec 03 10:57:12 crc kubenswrapper[4646]: I1203 10:57:12.740066 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:57:12 crc kubenswrapper[4646]: I1203 10:57:12.829180 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:57:12 crc kubenswrapper[4646]: I1203 10:57:12.902443 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clt4h\" (UniqueName: \"kubernetes.io/projected/e253153a-bb9a-4324-86ba-6fd879a3fff9-kube-api-access-clt4h\") pod \"e253153a-bb9a-4324-86ba-6fd879a3fff9\" (UID: \"e253153a-bb9a-4324-86ba-6fd879a3fff9\") " Dec 03 10:57:12 crc kubenswrapper[4646]: I1203 10:57:12.902528 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e253153a-bb9a-4324-86ba-6fd879a3fff9-catalog-content\") pod \"e253153a-bb9a-4324-86ba-6fd879a3fff9\" (UID: \"e253153a-bb9a-4324-86ba-6fd879a3fff9\") " Dec 03 10:57:12 crc kubenswrapper[4646]: I1203 10:57:12.902612 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e253153a-bb9a-4324-86ba-6fd879a3fff9-utilities\") pod \"e253153a-bb9a-4324-86ba-6fd879a3fff9\" (UID: \"e253153a-bb9a-4324-86ba-6fd879a3fff9\") " Dec 03 10:57:12 crc kubenswrapper[4646]: I1203 10:57:12.903594 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e253153a-bb9a-4324-86ba-6fd879a3fff9-utilities" (OuterVolumeSpecName: "utilities") pod "e253153a-bb9a-4324-86ba-6fd879a3fff9" (UID: "e253153a-bb9a-4324-86ba-6fd879a3fff9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:57:12 crc kubenswrapper[4646]: I1203 10:57:12.908158 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e253153a-bb9a-4324-86ba-6fd879a3fff9-kube-api-access-clt4h" (OuterVolumeSpecName: "kube-api-access-clt4h") pod "e253153a-bb9a-4324-86ba-6fd879a3fff9" (UID: "e253153a-bb9a-4324-86ba-6fd879a3fff9"). InnerVolumeSpecName "kube-api-access-clt4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.003366 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4048178-b02e-472c-a4db-5520e198217c-catalog-content\") pod \"d4048178-b02e-472c-a4db-5520e198217c\" (UID: \"d4048178-b02e-472c-a4db-5520e198217c\") " Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.003460 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crlwl\" (UniqueName: \"kubernetes.io/projected/d4048178-b02e-472c-a4db-5520e198217c-kube-api-access-crlwl\") pod \"d4048178-b02e-472c-a4db-5520e198217c\" (UID: \"d4048178-b02e-472c-a4db-5520e198217c\") " Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.003486 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4048178-b02e-472c-a4db-5520e198217c-utilities\") pod \"d4048178-b02e-472c-a4db-5520e198217c\" (UID: \"d4048178-b02e-472c-a4db-5520e198217c\") " Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.003765 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e253153a-bb9a-4324-86ba-6fd879a3fff9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.003778 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clt4h\" (UniqueName: \"kubernetes.io/projected/e253153a-bb9a-4324-86ba-6fd879a3fff9-kube-api-access-clt4h\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.005216 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4048178-b02e-472c-a4db-5520e198217c-utilities" (OuterVolumeSpecName: "utilities") pod "d4048178-b02e-472c-a4db-5520e198217c" (UID: "d4048178-b02e-472c-a4db-5520e198217c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.008403 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4048178-b02e-472c-a4db-5520e198217c-kube-api-access-crlwl" (OuterVolumeSpecName: "kube-api-access-crlwl") pod "d4048178-b02e-472c-a4db-5520e198217c" (UID: "d4048178-b02e-472c-a4db-5520e198217c"). InnerVolumeSpecName "kube-api-access-crlwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.026652 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4048178-b02e-472c-a4db-5520e198217c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4048178-b02e-472c-a4db-5520e198217c" (UID: "d4048178-b02e-472c-a4db-5520e198217c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.032089 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e253153a-bb9a-4324-86ba-6fd879a3fff9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e253153a-bb9a-4324-86ba-6fd879a3fff9" (UID: "e253153a-bb9a-4324-86ba-6fd879a3fff9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.105162 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4048178-b02e-472c-a4db-5520e198217c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.105208 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e253153a-bb9a-4324-86ba-6fd879a3fff9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.105224 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crlwl\" (UniqueName: \"kubernetes.io/projected/d4048178-b02e-472c-a4db-5520e198217c-kube-api-access-crlwl\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.105239 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4048178-b02e-472c-a4db-5520e198217c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.460586 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" podUID="7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" containerName="oauth-openshift" containerID="cri-o://41f8ff013e88e5457e4812f84d955623880c35a81841c74e9cb7e20a190bbb68" gracePeriod=15 Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.670785 4646 generic.go:334] "Generic (PLEG): container finished" podID="7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" containerID="41f8ff013e88e5457e4812f84d955623880c35a81841c74e9cb7e20a190bbb68" exitCode=0 Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.670852 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" event={"ID":"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924","Type":"ContainerDied","Data":"41f8ff013e88e5457e4812f84d955623880c35a81841c74e9cb7e20a190bbb68"} Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.673065 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlx92" event={"ID":"e253153a-bb9a-4324-86ba-6fd879a3fff9","Type":"ContainerDied","Data":"7f73e2f7e25a6fa23284eae31e7dd69b6be33cdd32093ee13ae1f954c08c975c"} Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.673117 4646 scope.go:117] "RemoveContainer" containerID="18980927df10a0f42dcdff7fd9355ed665533165f0646eff05a8e4e75217ab18" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.673243 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vlx92" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.681475 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ft6s8" event={"ID":"d4048178-b02e-472c-a4db-5520e198217c","Type":"ContainerDied","Data":"a1e687a7af51e853e36f4259d0e389a02233f3351e45a99850f0f6e8b842f7cf"} Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.681525 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ft6s8" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.711325 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vlx92"] Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.718668 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vlx92"] Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.737231 4646 scope.go:117] "RemoveContainer" containerID="a53a3e6f143876296e2cb5b04605a8f873b8e93540fd34ef91c7662bb747b9ec" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.754251 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ft6s8"] Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.758665 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ft6s8"] Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.772839 4646 scope.go:117] "RemoveContainer" containerID="d28ef5e9644f869e00ceee30738fce53a9df58c5ec833668550e7b5964312962" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.796535 4646 scope.go:117] "RemoveContainer" containerID="8ba10de4e74a0170bb5411d0efa011e0723e1ed26d850f475e9dc95c93f8562e" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.811086 4646 scope.go:117] "RemoveContainer" containerID="8079cf5ed9c0c19029e29c8582d726e72791785d01406d469278505e9f35daae" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.839021 4646 scope.go:117] "RemoveContainer" containerID="467ab9ea48488d24745018335241c53a98ef25a0ad8f6320ec1caf276b39d82f" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.840551 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.860351 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4048178-b02e-472c-a4db-5520e198217c" path="/var/lib/kubelet/pods/d4048178-b02e-472c-a4db-5520e198217c/volumes" Dec 03 10:57:13 crc kubenswrapper[4646]: I1203 10:57:13.861170 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e253153a-bb9a-4324-86ba-6fd879a3fff9" path="/var/lib/kubelet/pods/e253153a-bb9a-4324-86ba-6fd879a3fff9/volumes" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.019371 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-router-certs\") pod \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.019423 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-serving-cert\") pod \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.019455 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-session\") pod \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.019471 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-audit-dir\") pod \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.019490 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-audit-policies\") pod \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.019510 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-ocp-branding-template\") pod \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.019526 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-login\") pod \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.019562 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-cliconfig\") pod \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.019590 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-service-ca\") pod \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.019614 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-trusted-ca-bundle\") pod \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.019647 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-error\") pod \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.019669 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9jsx\" (UniqueName: \"kubernetes.io/projected/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-kube-api-access-m9jsx\") pod \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.019694 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-idp-0-file-data\") pod \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.019751 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-provider-selection\") pod \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\" (UID: \"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924\") " Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.021858 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" (UID: "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.022321 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" (UID: "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.022357 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" (UID: "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.022634 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" (UID: "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.023500 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" (UID: "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.025048 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" (UID: "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.025277 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" (UID: "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.025486 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" (UID: "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.028192 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" (UID: "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.028447 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" (UID: "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.029017 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" (UID: "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.032513 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" (UID: "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.032874 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-kube-api-access-m9jsx" (OuterVolumeSpecName: "kube-api-access-m9jsx") pod "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" (UID: "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924"). InnerVolumeSpecName "kube-api-access-m9jsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.033351 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" (UID: "7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.121500 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.121545 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.121560 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.121572 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.121583 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9jsx\" (UniqueName: \"kubernetes.io/projected/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-kube-api-access-m9jsx\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.121595 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.121615 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.121628 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.121639 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.121649 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.121661 4646 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.121672 4646 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.121682 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.121693 4646 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.690507 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" event={"ID":"7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924","Type":"ContainerDied","Data":"622bc5cf6198552c568236318014324bcba07c598f9e6704958cb17521ce03c9"} Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.690572 4646 scope.go:117] "RemoveContainer" containerID="41f8ff013e88e5457e4812f84d955623880c35a81841c74e9cb7e20a190bbb68" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.690670 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nb6wv" Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.717985 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nb6wv"] Dec 03 10:57:14 crc kubenswrapper[4646]: I1203 10:57:14.724560 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nb6wv"] Dec 03 10:57:15 crc kubenswrapper[4646]: I1203 10:57:15.791528 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:57:15 crc kubenswrapper[4646]: I1203 10:57:15.854410 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" path="/var/lib/kubelet/pods/7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924/volumes" Dec 03 10:57:15 crc kubenswrapper[4646]: I1203 10:57:15.970741 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:57:16 crc kubenswrapper[4646]: I1203 10:57:16.281937 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.674295 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5b64f96d58-hjfhh"] Dec 03 10:57:17 crc kubenswrapper[4646]: E1203 10:57:17.674747 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e253153a-bb9a-4324-86ba-6fd879a3fff9" containerName="extract-content" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.674776 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="e253153a-bb9a-4324-86ba-6fd879a3fff9" containerName="extract-content" Dec 03 10:57:17 crc kubenswrapper[4646]: E1203 10:57:17.674795 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702787d9-af42-442e-83eb-9b05c8b7caf8" containerName="pruner" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.674811 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="702787d9-af42-442e-83eb-9b05c8b7caf8" containerName="pruner" Dec 03 10:57:17 crc kubenswrapper[4646]: E1203 10:57:17.674834 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4048178-b02e-472c-a4db-5520e198217c" containerName="registry-server" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.674851 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4048178-b02e-472c-a4db-5520e198217c" containerName="registry-server" Dec 03 10:57:17 crc kubenswrapper[4646]: E1203 10:57:17.674872 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e253153a-bb9a-4324-86ba-6fd879a3fff9" containerName="registry-server" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.674888 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="e253153a-bb9a-4324-86ba-6fd879a3fff9" containerName="registry-server" Dec 03 10:57:17 crc kubenswrapper[4646]: E1203 10:57:17.674930 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4048178-b02e-472c-a4db-5520e198217c" containerName="extract-content" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.674977 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4048178-b02e-472c-a4db-5520e198217c" containerName="extract-content" Dec 03 10:57:17 crc kubenswrapper[4646]: E1203 10:57:17.675000 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4048178-b02e-472c-a4db-5520e198217c" containerName="extract-utilities" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.675017 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4048178-b02e-472c-a4db-5520e198217c" containerName="extract-utilities" Dec 03 10:57:17 crc kubenswrapper[4646]: E1203 10:57:17.675040 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e253153a-bb9a-4324-86ba-6fd879a3fff9" containerName="extract-utilities" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.675056 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="e253153a-bb9a-4324-86ba-6fd879a3fff9" containerName="extract-utilities" Dec 03 10:57:17 crc kubenswrapper[4646]: E1203 10:57:17.675081 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" containerName="oauth-openshift" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.675097 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" containerName="oauth-openshift" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.675326 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="e253153a-bb9a-4324-86ba-6fd879a3fff9" containerName="registry-server" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.675394 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="702787d9-af42-442e-83eb-9b05c8b7caf8" containerName="pruner" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.675425 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4048178-b02e-472c-a4db-5520e198217c" containerName="registry-server" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.675456 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f55d1b0-42a2-4c5e-8a6b-2b1fd2073924" containerName="oauth-openshift" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.676138 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.681703 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.682288 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.684240 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.685013 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.685991 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.686577 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.686634 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.686695 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.686850 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.687054 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.687124 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.698890 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.701452 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.710907 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.723953 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5b64f96d58-hjfhh"] Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.724485 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.769204 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-session\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.769319 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.769401 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1bb3191b-b967-4f9e-998b-2f77a283b25a-audit-dir\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.769436 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.769469 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-service-ca\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.769549 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.769622 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.769665 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-user-template-error\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.769726 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.769778 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-router-certs\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.769823 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1bb3191b-b967-4f9e-998b-2f77a283b25a-audit-policies\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.769886 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-user-template-login\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.769927 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.769979 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msbh2\" (UniqueName: \"kubernetes.io/projected/1bb3191b-b967-4f9e-998b-2f77a283b25a-kube-api-access-msbh2\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.871448 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msbh2\" (UniqueName: \"kubernetes.io/projected/1bb3191b-b967-4f9e-998b-2f77a283b25a-kube-api-access-msbh2\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.871512 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-session\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.871556 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.871588 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1bb3191b-b967-4f9e-998b-2f77a283b25a-audit-dir\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.871611 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-service-ca\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.871640 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.871705 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.871764 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.871804 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-user-template-error\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.871836 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.871871 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-router-certs\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.871900 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1bb3191b-b967-4f9e-998b-2f77a283b25a-audit-policies\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.871930 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-user-template-login\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.871962 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.873053 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1bb3191b-b967-4f9e-998b-2f77a283b25a-audit-dir\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.873285 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.874027 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-service-ca\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.874078 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1bb3191b-b967-4f9e-998b-2f77a283b25a-audit-policies\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.875297 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.880416 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-router-certs\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.880854 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-user-template-error\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.881088 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.881943 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-user-template-login\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.887977 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.890060 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-session\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.890161 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.894146 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msbh2\" (UniqueName: \"kubernetes.io/projected/1bb3191b-b967-4f9e-998b-2f77a283b25a-kube-api-access-msbh2\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:17 crc kubenswrapper[4646]: I1203 10:57:17.895810 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1bb3191b-b967-4f9e-998b-2f77a283b25a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5b64f96d58-hjfhh\" (UID: \"1bb3191b-b967-4f9e-998b-2f77a283b25a\") " pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:18 crc kubenswrapper[4646]: I1203 10:57:18.008760 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:18 crc kubenswrapper[4646]: I1203 10:57:18.419714 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5b64f96d58-hjfhh"] Dec 03 10:57:18 crc kubenswrapper[4646]: W1203 10:57:18.425623 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bb3191b_b967_4f9e_998b_2f77a283b25a.slice/crio-b608907ce51789604490530075c263ceef457a4471a2900cebd2dc4c094884b7 WatchSource:0}: Error finding container b608907ce51789604490530075c263ceef457a4471a2900cebd2dc4c094884b7: Status 404 returned error can't find the container with id b608907ce51789604490530075c263ceef457a4471a2900cebd2dc4c094884b7 Dec 03 10:57:18 crc kubenswrapper[4646]: I1203 10:57:18.722139 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" event={"ID":"1bb3191b-b967-4f9e-998b-2f77a283b25a","Type":"ContainerStarted","Data":"b608907ce51789604490530075c263ceef457a4471a2900cebd2dc4c094884b7"} Dec 03 10:57:18 crc kubenswrapper[4646]: I1203 10:57:18.792363 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rzw7n"] Dec 03 10:57:18 crc kubenswrapper[4646]: I1203 10:57:18.792717 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rzw7n" podUID="92bfb497-1e60-4973-a76c-0c686689cc53" containerName="registry-server" containerID="cri-o://25b4229b499c3a463ea2f7438e29c2e548681dc050d343482507ff659f409d26" gracePeriod=2 Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.703479 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.729314 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" event={"ID":"1bb3191b-b967-4f9e-998b-2f77a283b25a","Type":"ContainerStarted","Data":"258f8b64f69680d4f8d29eeb4e44c6d61e10e11c2acef85bfccd336b4bbb29b5"} Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.729692 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.733081 4646 generic.go:334] "Generic (PLEG): container finished" podID="92bfb497-1e60-4973-a76c-0c686689cc53" containerID="25b4229b499c3a463ea2f7438e29c2e548681dc050d343482507ff659f409d26" exitCode=0 Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.733116 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzw7n" event={"ID":"92bfb497-1e60-4973-a76c-0c686689cc53","Type":"ContainerDied","Data":"25b4229b499c3a463ea2f7438e29c2e548681dc050d343482507ff659f409d26"} Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.733140 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzw7n" event={"ID":"92bfb497-1e60-4973-a76c-0c686689cc53","Type":"ContainerDied","Data":"134b6d27fa65c504636837c4441c8d5eb488079d070ffe838b86441d2e24dcd6"} Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.733158 4646 scope.go:117] "RemoveContainer" containerID="25b4229b499c3a463ea2f7438e29c2e548681dc050d343482507ff659f409d26" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.733163 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rzw7n" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.738365 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.747959 4646 scope.go:117] "RemoveContainer" containerID="a4af1c9e511b645b729218dee2207679175801e838516dfc95116699413940ef" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.753713 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5b64f96d58-hjfhh" podStartSLOduration=31.753697284 podStartE2EDuration="31.753697284s" podCreationTimestamp="2025-12-03 10:56:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:57:19.751425359 +0000 UTC m=+216.214481494" watchObservedRunningTime="2025-12-03 10:57:19.753697284 +0000 UTC m=+216.216753419" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.764287 4646 scope.go:117] "RemoveContainer" containerID="4cefd6c44c3970872f8319f1ac85c862a6715a5836dd6cfb7c4eee8a6859a735" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.792540 4646 scope.go:117] "RemoveContainer" containerID="25b4229b499c3a463ea2f7438e29c2e548681dc050d343482507ff659f409d26" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.794991 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kv72s"] Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.795255 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kv72s" podUID="2b646c66-7457-42fe-abe5-0def6d0d3262" containerName="registry-server" containerID="cri-o://bbfc33fc24f85c44898bf009cb81a717db6413fdb2ff6f937c1b0ef35b54fed0" gracePeriod=2 Dec 03 10:57:19 crc kubenswrapper[4646]: E1203 10:57:19.798584 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25b4229b499c3a463ea2f7438e29c2e548681dc050d343482507ff659f409d26\": container with ID starting with 25b4229b499c3a463ea2f7438e29c2e548681dc050d343482507ff659f409d26 not found: ID does not exist" containerID="25b4229b499c3a463ea2f7438e29c2e548681dc050d343482507ff659f409d26" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.798624 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25b4229b499c3a463ea2f7438e29c2e548681dc050d343482507ff659f409d26"} err="failed to get container status \"25b4229b499c3a463ea2f7438e29c2e548681dc050d343482507ff659f409d26\": rpc error: code = NotFound desc = could not find container \"25b4229b499c3a463ea2f7438e29c2e548681dc050d343482507ff659f409d26\": container with ID starting with 25b4229b499c3a463ea2f7438e29c2e548681dc050d343482507ff659f409d26 not found: ID does not exist" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.798672 4646 scope.go:117] "RemoveContainer" containerID="a4af1c9e511b645b729218dee2207679175801e838516dfc95116699413940ef" Dec 03 10:57:19 crc kubenswrapper[4646]: E1203 10:57:19.799942 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4af1c9e511b645b729218dee2207679175801e838516dfc95116699413940ef\": container with ID starting with a4af1c9e511b645b729218dee2207679175801e838516dfc95116699413940ef not found: ID does not exist" containerID="a4af1c9e511b645b729218dee2207679175801e838516dfc95116699413940ef" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.799984 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4af1c9e511b645b729218dee2207679175801e838516dfc95116699413940ef"} err="failed to get container status \"a4af1c9e511b645b729218dee2207679175801e838516dfc95116699413940ef\": rpc error: code = NotFound desc = could not find container \"a4af1c9e511b645b729218dee2207679175801e838516dfc95116699413940ef\": container with ID starting with a4af1c9e511b645b729218dee2207679175801e838516dfc95116699413940ef not found: ID does not exist" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.800009 4646 scope.go:117] "RemoveContainer" containerID="4cefd6c44c3970872f8319f1ac85c862a6715a5836dd6cfb7c4eee8a6859a735" Dec 03 10:57:19 crc kubenswrapper[4646]: E1203 10:57:19.800589 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cefd6c44c3970872f8319f1ac85c862a6715a5836dd6cfb7c4eee8a6859a735\": container with ID starting with 4cefd6c44c3970872f8319f1ac85c862a6715a5836dd6cfb7c4eee8a6859a735 not found: ID does not exist" containerID="4cefd6c44c3970872f8319f1ac85c862a6715a5836dd6cfb7c4eee8a6859a735" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.800609 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cefd6c44c3970872f8319f1ac85c862a6715a5836dd6cfb7c4eee8a6859a735"} err="failed to get container status \"4cefd6c44c3970872f8319f1ac85c862a6715a5836dd6cfb7c4eee8a6859a735\": rpc error: code = NotFound desc = could not find container \"4cefd6c44c3970872f8319f1ac85c862a6715a5836dd6cfb7c4eee8a6859a735\": container with ID starting with 4cefd6c44c3970872f8319f1ac85c862a6715a5836dd6cfb7c4eee8a6859a735 not found: ID does not exist" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.896421 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6vcv\" (UniqueName: \"kubernetes.io/projected/92bfb497-1e60-4973-a76c-0c686689cc53-kube-api-access-s6vcv\") pod \"92bfb497-1e60-4973-a76c-0c686689cc53\" (UID: \"92bfb497-1e60-4973-a76c-0c686689cc53\") " Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.896539 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92bfb497-1e60-4973-a76c-0c686689cc53-catalog-content\") pod \"92bfb497-1e60-4973-a76c-0c686689cc53\" (UID: \"92bfb497-1e60-4973-a76c-0c686689cc53\") " Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.896571 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92bfb497-1e60-4973-a76c-0c686689cc53-utilities\") pod \"92bfb497-1e60-4973-a76c-0c686689cc53\" (UID: \"92bfb497-1e60-4973-a76c-0c686689cc53\") " Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.899763 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92bfb497-1e60-4973-a76c-0c686689cc53-utilities" (OuterVolumeSpecName: "utilities") pod "92bfb497-1e60-4973-a76c-0c686689cc53" (UID: "92bfb497-1e60-4973-a76c-0c686689cc53"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.917021 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92bfb497-1e60-4973-a76c-0c686689cc53-kube-api-access-s6vcv" (OuterVolumeSpecName: "kube-api-access-s6vcv") pod "92bfb497-1e60-4973-a76c-0c686689cc53" (UID: "92bfb497-1e60-4973-a76c-0c686689cc53"). InnerVolumeSpecName "kube-api-access-s6vcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.993688 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92bfb497-1e60-4973-a76c-0c686689cc53-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92bfb497-1e60-4973-a76c-0c686689cc53" (UID: "92bfb497-1e60-4973-a76c-0c686689cc53"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.997910 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92bfb497-1e60-4973-a76c-0c686689cc53-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.997948 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92bfb497-1e60-4973-a76c-0c686689cc53-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:19 crc kubenswrapper[4646]: I1203 10:57:19.997957 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6vcv\" (UniqueName: \"kubernetes.io/projected/92bfb497-1e60-4973-a76c-0c686689cc53-kube-api-access-s6vcv\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.065324 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rzw7n"] Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.075409 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rzw7n"] Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.235978 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.401194 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxktn\" (UniqueName: \"kubernetes.io/projected/2b646c66-7457-42fe-abe5-0def6d0d3262-kube-api-access-rxktn\") pod \"2b646c66-7457-42fe-abe5-0def6d0d3262\" (UID: \"2b646c66-7457-42fe-abe5-0def6d0d3262\") " Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.401249 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b646c66-7457-42fe-abe5-0def6d0d3262-catalog-content\") pod \"2b646c66-7457-42fe-abe5-0def6d0d3262\" (UID: \"2b646c66-7457-42fe-abe5-0def6d0d3262\") " Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.401281 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b646c66-7457-42fe-abe5-0def6d0d3262-utilities\") pod \"2b646c66-7457-42fe-abe5-0def6d0d3262\" (UID: \"2b646c66-7457-42fe-abe5-0def6d0d3262\") " Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.402016 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b646c66-7457-42fe-abe5-0def6d0d3262-utilities" (OuterVolumeSpecName: "utilities") pod "2b646c66-7457-42fe-abe5-0def6d0d3262" (UID: "2b646c66-7457-42fe-abe5-0def6d0d3262"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.404324 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b646c66-7457-42fe-abe5-0def6d0d3262-kube-api-access-rxktn" (OuterVolumeSpecName: "kube-api-access-rxktn") pod "2b646c66-7457-42fe-abe5-0def6d0d3262" (UID: "2b646c66-7457-42fe-abe5-0def6d0d3262"). InnerVolumeSpecName "kube-api-access-rxktn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.447053 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b646c66-7457-42fe-abe5-0def6d0d3262-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b646c66-7457-42fe-abe5-0def6d0d3262" (UID: "2b646c66-7457-42fe-abe5-0def6d0d3262"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.503155 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxktn\" (UniqueName: \"kubernetes.io/projected/2b646c66-7457-42fe-abe5-0def6d0d3262-kube-api-access-rxktn\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.503191 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b646c66-7457-42fe-abe5-0def6d0d3262-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.503204 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b646c66-7457-42fe-abe5-0def6d0d3262-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.741214 4646 generic.go:334] "Generic (PLEG): container finished" podID="2b646c66-7457-42fe-abe5-0def6d0d3262" containerID="bbfc33fc24f85c44898bf009cb81a717db6413fdb2ff6f937c1b0ef35b54fed0" exitCode=0 Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.741256 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kv72s" event={"ID":"2b646c66-7457-42fe-abe5-0def6d0d3262","Type":"ContainerDied","Data":"bbfc33fc24f85c44898bf009cb81a717db6413fdb2ff6f937c1b0ef35b54fed0"} Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.741325 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kv72s" event={"ID":"2b646c66-7457-42fe-abe5-0def6d0d3262","Type":"ContainerDied","Data":"eed51c0a4ca3ceeed1093d11420d8f781db8e0d56b0cff6dedab7d72141e87ae"} Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.741269 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kv72s" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.741363 4646 scope.go:117] "RemoveContainer" containerID="bbfc33fc24f85c44898bf009cb81a717db6413fdb2ff6f937c1b0ef35b54fed0" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.756647 4646 scope.go:117] "RemoveContainer" containerID="3da029f6f0eeaff30acf62ee29ad9da10abc22622f09927720d03d14b88cf5dc" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.765524 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kv72s"] Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.784504 4646 scope.go:117] "RemoveContainer" containerID="492ce31ba6978193f06ecdb65ff05603d5e24ea781aa7ac2d212f3902afba3f0" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.785473 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kv72s"] Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.803877 4646 scope.go:117] "RemoveContainer" containerID="bbfc33fc24f85c44898bf009cb81a717db6413fdb2ff6f937c1b0ef35b54fed0" Dec 03 10:57:20 crc kubenswrapper[4646]: E1203 10:57:20.804378 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbfc33fc24f85c44898bf009cb81a717db6413fdb2ff6f937c1b0ef35b54fed0\": container with ID starting with bbfc33fc24f85c44898bf009cb81a717db6413fdb2ff6f937c1b0ef35b54fed0 not found: ID does not exist" containerID="bbfc33fc24f85c44898bf009cb81a717db6413fdb2ff6f937c1b0ef35b54fed0" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.804409 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbfc33fc24f85c44898bf009cb81a717db6413fdb2ff6f937c1b0ef35b54fed0"} err="failed to get container status \"bbfc33fc24f85c44898bf009cb81a717db6413fdb2ff6f937c1b0ef35b54fed0\": rpc error: code = NotFound desc = could not find container \"bbfc33fc24f85c44898bf009cb81a717db6413fdb2ff6f937c1b0ef35b54fed0\": container with ID starting with bbfc33fc24f85c44898bf009cb81a717db6413fdb2ff6f937c1b0ef35b54fed0 not found: ID does not exist" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.804435 4646 scope.go:117] "RemoveContainer" containerID="3da029f6f0eeaff30acf62ee29ad9da10abc22622f09927720d03d14b88cf5dc" Dec 03 10:57:20 crc kubenswrapper[4646]: E1203 10:57:20.804918 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3da029f6f0eeaff30acf62ee29ad9da10abc22622f09927720d03d14b88cf5dc\": container with ID starting with 3da029f6f0eeaff30acf62ee29ad9da10abc22622f09927720d03d14b88cf5dc not found: ID does not exist" containerID="3da029f6f0eeaff30acf62ee29ad9da10abc22622f09927720d03d14b88cf5dc" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.804957 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3da029f6f0eeaff30acf62ee29ad9da10abc22622f09927720d03d14b88cf5dc"} err="failed to get container status \"3da029f6f0eeaff30acf62ee29ad9da10abc22622f09927720d03d14b88cf5dc\": rpc error: code = NotFound desc = could not find container \"3da029f6f0eeaff30acf62ee29ad9da10abc22622f09927720d03d14b88cf5dc\": container with ID starting with 3da029f6f0eeaff30acf62ee29ad9da10abc22622f09927720d03d14b88cf5dc not found: ID does not exist" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.804997 4646 scope.go:117] "RemoveContainer" containerID="492ce31ba6978193f06ecdb65ff05603d5e24ea781aa7ac2d212f3902afba3f0" Dec 03 10:57:20 crc kubenswrapper[4646]: E1203 10:57:20.805893 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"492ce31ba6978193f06ecdb65ff05603d5e24ea781aa7ac2d212f3902afba3f0\": container with ID starting with 492ce31ba6978193f06ecdb65ff05603d5e24ea781aa7ac2d212f3902afba3f0 not found: ID does not exist" containerID="492ce31ba6978193f06ecdb65ff05603d5e24ea781aa7ac2d212f3902afba3f0" Dec 03 10:57:20 crc kubenswrapper[4646]: I1203 10:57:20.805917 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"492ce31ba6978193f06ecdb65ff05603d5e24ea781aa7ac2d212f3902afba3f0"} err="failed to get container status \"492ce31ba6978193f06ecdb65ff05603d5e24ea781aa7ac2d212f3902afba3f0\": rpc error: code = NotFound desc = could not find container \"492ce31ba6978193f06ecdb65ff05603d5e24ea781aa7ac2d212f3902afba3f0\": container with ID starting with 492ce31ba6978193f06ecdb65ff05603d5e24ea781aa7ac2d212f3902afba3f0 not found: ID does not exist" Dec 03 10:57:21 crc kubenswrapper[4646]: I1203 10:57:21.855083 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b646c66-7457-42fe-abe5-0def6d0d3262" path="/var/lib/kubelet/pods/2b646c66-7457-42fe-abe5-0def6d0d3262/volumes" Dec 03 10:57:21 crc kubenswrapper[4646]: I1203 10:57:21.856675 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92bfb497-1e60-4973-a76c-0c686689cc53" path="/var/lib/kubelet/pods/92bfb497-1e60-4973-a76c-0c686689cc53/volumes" Dec 03 10:57:25 crc kubenswrapper[4646]: I1203 10:57:25.964713 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:57:25 crc kubenswrapper[4646]: I1203 10:57:25.965252 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:57:25 crc kubenswrapper[4646]: I1203 10:57:25.965310 4646 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 10:57:25 crc kubenswrapper[4646]: I1203 10:57:25.965906 4646 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441"} pod="openshift-machine-config-operator/machine-config-daemon-pndzt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 10:57:25 crc kubenswrapper[4646]: I1203 10:57:25.965959 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" containerID="cri-o://b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441" gracePeriod=600 Dec 03 10:57:26 crc kubenswrapper[4646]: I1203 10:57:26.773317 4646 generic.go:334] "Generic (PLEG): container finished" podID="9695d947-fc43-4625-86af-a1a3e177bb27" containerID="b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441" exitCode=0 Dec 03 10:57:26 crc kubenswrapper[4646]: I1203 10:57:26.773433 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerDied","Data":"b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441"} Dec 03 10:57:27 crc kubenswrapper[4646]: I1203 10:57:27.785273 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerStarted","Data":"18f4f8a8e5b1f28504a2c74c33b0ac967e695b60d0b4a11af354622575d39b93"} Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.036161 4646 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.037062 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92bfb497-1e60-4973-a76c-0c686689cc53" containerName="extract-content" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.037079 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="92bfb497-1e60-4973-a76c-0c686689cc53" containerName="extract-content" Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.037095 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92bfb497-1e60-4973-a76c-0c686689cc53" containerName="extract-utilities" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.037102 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="92bfb497-1e60-4973-a76c-0c686689cc53" containerName="extract-utilities" Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.037113 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b646c66-7457-42fe-abe5-0def6d0d3262" containerName="extract-utilities" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.037121 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b646c66-7457-42fe-abe5-0def6d0d3262" containerName="extract-utilities" Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.037138 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92bfb497-1e60-4973-a76c-0c686689cc53" containerName="registry-server" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.037145 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="92bfb497-1e60-4973-a76c-0c686689cc53" containerName="registry-server" Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.037162 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b646c66-7457-42fe-abe5-0def6d0d3262" containerName="registry-server" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.037170 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b646c66-7457-42fe-abe5-0def6d0d3262" containerName="registry-server" Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.037223 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b646c66-7457-42fe-abe5-0def6d0d3262" containerName="extract-content" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.037231 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b646c66-7457-42fe-abe5-0def6d0d3262" containerName="extract-content" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.037372 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b646c66-7457-42fe-abe5-0def6d0d3262" containerName="registry-server" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.037392 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="92bfb497-1e60-4973-a76c-0c686689cc53" containerName="registry-server" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.037825 4646 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.037935 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.038108 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd" gracePeriod=15 Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.038670 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943" gracePeriod=15 Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.038655 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde" gracePeriod=15 Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.038810 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708" gracePeriod=15 Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.038849 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d" gracePeriod=15 Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.041763 4646 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.044917 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.044946 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.044961 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.044972 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.044991 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.045048 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.045074 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.045086 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.045103 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.045641 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.045669 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.045682 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.045700 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.045782 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.045993 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.046011 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.046026 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.046042 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.046054 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.046066 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.078762 4646 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.155:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.156040 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.156080 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.156104 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.156171 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.156207 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.156234 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.156252 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.156282 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.257407 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.257750 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.257772 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.257798 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.257778 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.257838 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.257852 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.257873 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.257891 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.257905 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.257964 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.257988 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.258009 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.258030 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.257540 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.258023 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.380103 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: W1203 10:57:40.408367 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-7fb2ecc15802d507628ac250004a3126546f5f75ed548258c8d32cac49a7ce63 WatchSource:0}: Error finding container 7fb2ecc15802d507628ac250004a3126546f5f75ed548258c8d32cac49a7ce63: Status 404 returned error can't find the container with id 7fb2ecc15802d507628ac250004a3126546f5f75ed548258c8d32cac49a7ce63 Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.416466 4646 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.155:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187daf6110187e58 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 10:57:40.41583164 +0000 UTC m=+236.878887815,LastTimestamp:2025-12-03 10:57:40.41583164 +0000 UTC m=+236.878887815,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.664927 4646 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.664984 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.861705 4646 generic.go:334] "Generic (PLEG): container finished" podID="9030be92-e97b-4760-a4cf-abe87c241ec6" containerID="2c5fbb57f828b7e822c0fbf9cce6aa353afa23a324bdad1c6b451d0de880d21a" exitCode=0 Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.861785 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9030be92-e97b-4760-a4cf-abe87c241ec6","Type":"ContainerDied","Data":"2c5fbb57f828b7e822c0fbf9cce6aa353afa23a324bdad1c6b451d0de880d21a"} Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.862576 4646 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.862940 4646 status_manager.go:851] "Failed to get status for pod" podUID="9030be92-e97b-4760-a4cf-abe87c241ec6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.865735 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.867176 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.868047 4646 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708" exitCode=0 Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.868070 4646 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde" exitCode=0 Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.868079 4646 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943" exitCode=0 Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.868087 4646 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d" exitCode=2 Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.868142 4646 scope.go:117] "RemoveContainer" containerID="a436f837b1a944bf714ca4fd9dda098fe9dbd261b6adf09239ecc98b1c86351d" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.869763 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"625a623d805ff19721ca78009ff358b83659042674a5fae0abec2554449a54d5"} Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.869813 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"7fb2ecc15802d507628ac250004a3126546f5f75ed548258c8d32cac49a7ce63"} Dec 03 10:57:40 crc kubenswrapper[4646]: E1203 10:57:40.870476 4646 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.155:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.870849 4646 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:40 crc kubenswrapper[4646]: I1203 10:57:40.871389 4646 status_manager.go:851] "Failed to get status for pod" podUID="9030be92-e97b-4760-a4cf-abe87c241ec6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:41 crc kubenswrapper[4646]: I1203 10:57:41.883523 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.289273 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.290176 4646 status_manager.go:851] "Failed to get status for pod" podUID="9030be92-e97b-4760-a4cf-abe87c241ec6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.390220 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9030be92-e97b-4760-a4cf-abe87c241ec6-kubelet-dir\") pod \"9030be92-e97b-4760-a4cf-abe87c241ec6\" (UID: \"9030be92-e97b-4760-a4cf-abe87c241ec6\") " Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.390317 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9030be92-e97b-4760-a4cf-abe87c241ec6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9030be92-e97b-4760-a4cf-abe87c241ec6" (UID: "9030be92-e97b-4760-a4cf-abe87c241ec6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.390326 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9030be92-e97b-4760-a4cf-abe87c241ec6-var-lock\") pod \"9030be92-e97b-4760-a4cf-abe87c241ec6\" (UID: \"9030be92-e97b-4760-a4cf-abe87c241ec6\") " Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.390417 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9030be92-e97b-4760-a4cf-abe87c241ec6-kube-api-access\") pod \"9030be92-e97b-4760-a4cf-abe87c241ec6\" (UID: \"9030be92-e97b-4760-a4cf-abe87c241ec6\") " Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.390480 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9030be92-e97b-4760-a4cf-abe87c241ec6-var-lock" (OuterVolumeSpecName: "var-lock") pod "9030be92-e97b-4760-a4cf-abe87c241ec6" (UID: "9030be92-e97b-4760-a4cf-abe87c241ec6"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.390654 4646 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9030be92-e97b-4760-a4cf-abe87c241ec6-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.390669 4646 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9030be92-e97b-4760-a4cf-abe87c241ec6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.398537 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9030be92-e97b-4760-a4cf-abe87c241ec6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9030be92-e97b-4760-a4cf-abe87c241ec6" (UID: "9030be92-e97b-4760-a4cf-abe87c241ec6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.401074 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.401657 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.402096 4646 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.402379 4646 status_manager.go:851] "Failed to get status for pod" podUID="9030be92-e97b-4760-a4cf-abe87c241ec6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.491966 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.492094 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.492139 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.492156 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.492177 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.492324 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.492623 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9030be92-e97b-4760-a4cf-abe87c241ec6-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.492652 4646 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.492669 4646 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.492685 4646 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.896124 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9030be92-e97b-4760-a4cf-abe87c241ec6","Type":"ContainerDied","Data":"3fce1dd1b54f4c11cf63f283ea671167c74d39e2f4821c842396ecf3afb00f19"} Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.896162 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.896187 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fce1dd1b54f4c11cf63f283ea671167c74d39e2f4821c842396ecf3afb00f19" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.902428 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.905002 4646 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd" exitCode=0 Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.905066 4646 scope.go:117] "RemoveContainer" containerID="6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.905251 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.938668 4646 scope.go:117] "RemoveContainer" containerID="1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.949261 4646 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.949718 4646 status_manager.go:851] "Failed to get status for pod" podUID="9030be92-e97b-4760-a4cf-abe87c241ec6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.950255 4646 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.950627 4646 status_manager.go:851] "Failed to get status for pod" podUID="9030be92-e97b-4760-a4cf-abe87c241ec6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.961769 4646 scope.go:117] "RemoveContainer" containerID="88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943" Dec 03 10:57:42 crc kubenswrapper[4646]: I1203 10:57:42.984634 4646 scope.go:117] "RemoveContainer" containerID="e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.008142 4646 scope.go:117] "RemoveContainer" containerID="458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.027055 4646 scope.go:117] "RemoveContainer" containerID="fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.051756 4646 scope.go:117] "RemoveContainer" containerID="6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708" Dec 03 10:57:43 crc kubenswrapper[4646]: E1203 10:57:43.052234 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\": container with ID starting with 6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708 not found: ID does not exist" containerID="6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.052283 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708"} err="failed to get container status \"6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\": rpc error: code = NotFound desc = could not find container \"6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708\": container with ID starting with 6d0946c54c379d408af4954a3b3d71d4e113b990a3799c3a08cf23ed83904708 not found: ID does not exist" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.052317 4646 scope.go:117] "RemoveContainer" containerID="1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde" Dec 03 10:57:43 crc kubenswrapper[4646]: E1203 10:57:43.053502 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\": container with ID starting with 1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde not found: ID does not exist" containerID="1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.053572 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde"} err="failed to get container status \"1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\": rpc error: code = NotFound desc = could not find container \"1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde\": container with ID starting with 1711402ad9cfbee7d896f1baf4527ff45941ad24f5a02a37fdf957e1fb91efde not found: ID does not exist" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.053604 4646 scope.go:117] "RemoveContainer" containerID="88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943" Dec 03 10:57:43 crc kubenswrapper[4646]: E1203 10:57:43.053955 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\": container with ID starting with 88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943 not found: ID does not exist" containerID="88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.053984 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943"} err="failed to get container status \"88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\": rpc error: code = NotFound desc = could not find container \"88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943\": container with ID starting with 88a546ca9256a60b6fe8710b9a51f622ce24b81343661c147c396e464da88943 not found: ID does not exist" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.054001 4646 scope.go:117] "RemoveContainer" containerID="e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d" Dec 03 10:57:43 crc kubenswrapper[4646]: E1203 10:57:43.054251 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\": container with ID starting with e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d not found: ID does not exist" containerID="e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.054280 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d"} err="failed to get container status \"e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\": rpc error: code = NotFound desc = could not find container \"e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d\": container with ID starting with e3377faa33702746dd2c4f7dc36d53dfe9ce59b3c5adc4dd0dbc735b8e03415d not found: ID does not exist" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.054298 4646 scope.go:117] "RemoveContainer" containerID="458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd" Dec 03 10:57:43 crc kubenswrapper[4646]: E1203 10:57:43.054703 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\": container with ID starting with 458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd not found: ID does not exist" containerID="458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.054748 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd"} err="failed to get container status \"458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\": rpc error: code = NotFound desc = could not find container \"458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd\": container with ID starting with 458dda396c6435607a90a0b811ed9542f5126531c6456076e6f79e0bd9a792bd not found: ID does not exist" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.054771 4646 scope.go:117] "RemoveContainer" containerID="fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0" Dec 03 10:57:43 crc kubenswrapper[4646]: E1203 10:57:43.055101 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\": container with ID starting with fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0 not found: ID does not exist" containerID="fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.055125 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0"} err="failed to get container status \"fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\": rpc error: code = NotFound desc = could not find container \"fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0\": container with ID starting with fe1982f361f72d9348e498bba5bc69804903341fdd453897ee0064d2591adec0 not found: ID does not exist" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.851445 4646 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.851981 4646 status_manager.go:851] "Failed to get status for pod" podUID="9030be92-e97b-4760-a4cf-abe87c241ec6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:43 crc kubenswrapper[4646]: I1203 10:57:43.854961 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 03 10:57:47 crc kubenswrapper[4646]: E1203 10:57:47.855625 4646 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.155:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187daf6110187e58 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 10:57:40.41583164 +0000 UTC m=+236.878887815,LastTimestamp:2025-12-03 10:57:40.41583164 +0000 UTC m=+236.878887815,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 10:57:48 crc kubenswrapper[4646]: E1203 10:57:48.751081 4646 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:48 crc kubenswrapper[4646]: E1203 10:57:48.751600 4646 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:48 crc kubenswrapper[4646]: E1203 10:57:48.752170 4646 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:48 crc kubenswrapper[4646]: E1203 10:57:48.752692 4646 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:48 crc kubenswrapper[4646]: E1203 10:57:48.753182 4646 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:48 crc kubenswrapper[4646]: I1203 10:57:48.753230 4646 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 03 10:57:48 crc kubenswrapper[4646]: E1203 10:57:48.753660 4646 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="200ms" Dec 03 10:57:48 crc kubenswrapper[4646]: E1203 10:57:48.954974 4646 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="400ms" Dec 03 10:57:49 crc kubenswrapper[4646]: E1203 10:57:49.356135 4646 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="800ms" Dec 03 10:57:50 crc kubenswrapper[4646]: E1203 10:57:50.158340 4646 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="1.6s" Dec 03 10:57:51 crc kubenswrapper[4646]: E1203 10:57:51.760022 4646 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="3.2s" Dec 03 10:57:53 crc kubenswrapper[4646]: I1203 10:57:53.853466 4646 status_manager.go:851] "Failed to get status for pod" podUID="9030be92-e97b-4760-a4cf-abe87c241ec6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:54 crc kubenswrapper[4646]: I1203 10:57:54.848134 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:54 crc kubenswrapper[4646]: I1203 10:57:54.849393 4646 status_manager.go:851] "Failed to get status for pod" podUID="9030be92-e97b-4760-a4cf-abe87c241ec6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:54 crc kubenswrapper[4646]: I1203 10:57:54.869078 4646 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8869765-e63a-44d3-912f-8db6a978db59" Dec 03 10:57:54 crc kubenswrapper[4646]: I1203 10:57:54.869115 4646 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8869765-e63a-44d3-912f-8db6a978db59" Dec 03 10:57:54 crc kubenswrapper[4646]: E1203 10:57:54.870132 4646 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:54 crc kubenswrapper[4646]: I1203 10:57:54.870992 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:54 crc kubenswrapper[4646]: W1203 10:57:54.901490 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-9760a5284adc61d4a16b5568ae1e5d103b807e5cb1315026546b723bfcb50563 WatchSource:0}: Error finding container 9760a5284adc61d4a16b5568ae1e5d103b807e5cb1315026546b723bfcb50563: Status 404 returned error can't find the container with id 9760a5284adc61d4a16b5568ae1e5d103b807e5cb1315026546b723bfcb50563 Dec 03 10:57:54 crc kubenswrapper[4646]: E1203 10:57:54.961303 4646 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.155:6443: connect: connection refused" interval="6.4s" Dec 03 10:57:54 crc kubenswrapper[4646]: I1203 10:57:54.978633 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 10:57:54 crc kubenswrapper[4646]: I1203 10:57:54.978738 4646 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e" exitCode=1 Dec 03 10:57:54 crc kubenswrapper[4646]: I1203 10:57:54.978827 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e"} Dec 03 10:57:54 crc kubenswrapper[4646]: I1203 10:57:54.979748 4646 scope.go:117] "RemoveContainer" containerID="29bc903171e0e42cc70b0a2dc405817a1b0d1b7a1109c297cec4999dbe32d08e" Dec 03 10:57:54 crc kubenswrapper[4646]: I1203 10:57:54.979799 4646 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:54 crc kubenswrapper[4646]: I1203 10:57:54.980194 4646 status_manager.go:851] "Failed to get status for pod" podUID="9030be92-e97b-4760-a4cf-abe87c241ec6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:54 crc kubenswrapper[4646]: I1203 10:57:54.981512 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9760a5284adc61d4a16b5568ae1e5d103b807e5cb1315026546b723bfcb50563"} Dec 03 10:57:55 crc kubenswrapper[4646]: I1203 10:57:55.990619 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 10:57:55 crc kubenswrapper[4646]: I1203 10:57:55.991105 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"25e29177747e20f13bb6a99c73a3e555a1520bd7cea059ca86eb7efa5d84d6ed"} Dec 03 10:57:55 crc kubenswrapper[4646]: I1203 10:57:55.991871 4646 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:55 crc kubenswrapper[4646]: I1203 10:57:55.992097 4646 status_manager.go:851] "Failed to get status for pod" podUID="9030be92-e97b-4760-a4cf-abe87c241ec6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:55 crc kubenswrapper[4646]: I1203 10:57:55.993449 4646 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="7cadfab4e8dec63660f0d394a0850ef0cef7895eb8367dc4adc9e05a4534126b" exitCode=0 Dec 03 10:57:55 crc kubenswrapper[4646]: I1203 10:57:55.993478 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"7cadfab4e8dec63660f0d394a0850ef0cef7895eb8367dc4adc9e05a4534126b"} Dec 03 10:57:55 crc kubenswrapper[4646]: I1203 10:57:55.993629 4646 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8869765-e63a-44d3-912f-8db6a978db59" Dec 03 10:57:55 crc kubenswrapper[4646]: I1203 10:57:55.993648 4646 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8869765-e63a-44d3-912f-8db6a978db59" Dec 03 10:57:55 crc kubenswrapper[4646]: E1203 10:57:55.994063 4646 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:55 crc kubenswrapper[4646]: I1203 10:57:55.994105 4646 status_manager.go:851] "Failed to get status for pod" podUID="9030be92-e97b-4760-a4cf-abe87c241ec6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:55 crc kubenswrapper[4646]: I1203 10:57:55.994663 4646 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.155:6443: connect: connection refused" Dec 03 10:57:56 crc kubenswrapper[4646]: I1203 10:57:56.980377 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:57:57 crc kubenswrapper[4646]: I1203 10:57:57.002597 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"edf2acb2f844e8d69ba9eac504f476218c5987e2abae6450fd1cc1068810a4f3"} Dec 03 10:57:57 crc kubenswrapper[4646]: I1203 10:57:57.002650 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ff125b51335d32cc99edbc2d366e12d345201a949050a37dfb64ae21452b3a89"} Dec 03 10:57:58 crc kubenswrapper[4646]: I1203 10:57:58.011281 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e229091e76a7a21b955ad0dab3a76259856eba599628744b8e6f74f2d16b6d43"} Dec 03 10:57:58 crc kubenswrapper[4646]: I1203 10:57:58.011678 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"db7e3359fe9ebf15f40cdae7ee800549a1a3bc854253d49ff1c0ce828ec39f72"} Dec 03 10:57:58 crc kubenswrapper[4646]: I1203 10:57:58.011694 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"465a5308360dcf8ad6ee30e5ad619c752c228f23d00ac5a7d28f710deaaa993c"} Dec 03 10:57:58 crc kubenswrapper[4646]: I1203 10:57:58.011713 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:58 crc kubenswrapper[4646]: I1203 10:57:58.011583 4646 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8869765-e63a-44d3-912f-8db6a978db59" Dec 03 10:57:58 crc kubenswrapper[4646]: I1203 10:57:58.011741 4646 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8869765-e63a-44d3-912f-8db6a978db59" Dec 03 10:57:58 crc kubenswrapper[4646]: I1203 10:57:58.595582 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:57:58 crc kubenswrapper[4646]: I1203 10:57:58.602326 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:57:59 crc kubenswrapper[4646]: I1203 10:57:59.871655 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:59 crc kubenswrapper[4646]: I1203 10:57:59.872488 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:57:59 crc kubenswrapper[4646]: I1203 10:57:59.878493 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:58:03 crc kubenswrapper[4646]: I1203 10:58:03.120629 4646 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:58:03 crc kubenswrapper[4646]: I1203 10:58:03.877616 4646 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="473e50f9-0dc2-45ff-ab8a-ee3c54b8960f" Dec 03 10:58:04 crc kubenswrapper[4646]: I1203 10:58:04.045663 4646 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8869765-e63a-44d3-912f-8db6a978db59" Dec 03 10:58:04 crc kubenswrapper[4646]: I1203 10:58:04.045691 4646 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8869765-e63a-44d3-912f-8db6a978db59" Dec 03 10:58:04 crc kubenswrapper[4646]: I1203 10:58:04.049319 4646 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="473e50f9-0dc2-45ff-ab8a-ee3c54b8960f" Dec 03 10:58:04 crc kubenswrapper[4646]: I1203 10:58:04.049978 4646 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://ff125b51335d32cc99edbc2d366e12d345201a949050a37dfb64ae21452b3a89" Dec 03 10:58:04 crc kubenswrapper[4646]: I1203 10:58:04.050006 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:58:05 crc kubenswrapper[4646]: I1203 10:58:05.050026 4646 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8869765-e63a-44d3-912f-8db6a978db59" Dec 03 10:58:05 crc kubenswrapper[4646]: I1203 10:58:05.050055 4646 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8869765-e63a-44d3-912f-8db6a978db59" Dec 03 10:58:05 crc kubenswrapper[4646]: I1203 10:58:05.053854 4646 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="473e50f9-0dc2-45ff-ab8a-ee3c54b8960f" Dec 03 10:58:06 crc kubenswrapper[4646]: I1203 10:58:06.989754 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 10:58:12 crc kubenswrapper[4646]: I1203 10:58:12.549094 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 10:58:12 crc kubenswrapper[4646]: I1203 10:58:12.806880 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 10:58:13 crc kubenswrapper[4646]: I1203 10:58:13.233690 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 10:58:13 crc kubenswrapper[4646]: I1203 10:58:13.521842 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 10:58:13 crc kubenswrapper[4646]: I1203 10:58:13.556090 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 10:58:13 crc kubenswrapper[4646]: I1203 10:58:13.695993 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 10:58:13 crc kubenswrapper[4646]: I1203 10:58:13.717972 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 10:58:13 crc kubenswrapper[4646]: I1203 10:58:13.896386 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 10:58:14 crc kubenswrapper[4646]: I1203 10:58:14.112491 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 10:58:14 crc kubenswrapper[4646]: I1203 10:58:14.113393 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 10:58:14 crc kubenswrapper[4646]: I1203 10:58:14.236683 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 10:58:14 crc kubenswrapper[4646]: I1203 10:58:14.250203 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 10:58:14 crc kubenswrapper[4646]: I1203 10:58:14.270195 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 10:58:14 crc kubenswrapper[4646]: I1203 10:58:14.352191 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 10:58:14 crc kubenswrapper[4646]: I1203 10:58:14.705687 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 10:58:14 crc kubenswrapper[4646]: I1203 10:58:14.818258 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 10:58:14 crc kubenswrapper[4646]: I1203 10:58:14.856435 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 10:58:14 crc kubenswrapper[4646]: I1203 10:58:14.914522 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 10:58:15 crc kubenswrapper[4646]: I1203 10:58:15.406304 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 10:58:15 crc kubenswrapper[4646]: I1203 10:58:15.701195 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 10:58:15 crc kubenswrapper[4646]: I1203 10:58:15.866300 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 10:58:15 crc kubenswrapper[4646]: I1203 10:58:15.948976 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 10:58:16 crc kubenswrapper[4646]: I1203 10:58:16.148892 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 10:58:16 crc kubenswrapper[4646]: I1203 10:58:16.175224 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 10:58:16 crc kubenswrapper[4646]: I1203 10:58:16.225585 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 10:58:16 crc kubenswrapper[4646]: I1203 10:58:16.284303 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 10:58:16 crc kubenswrapper[4646]: I1203 10:58:16.288921 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 10:58:16 crc kubenswrapper[4646]: I1203 10:58:16.337984 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 10:58:16 crc kubenswrapper[4646]: I1203 10:58:16.387278 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 10:58:16 crc kubenswrapper[4646]: I1203 10:58:16.390547 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 10:58:16 crc kubenswrapper[4646]: I1203 10:58:16.429408 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 10:58:16 crc kubenswrapper[4646]: I1203 10:58:16.484942 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 10:58:16 crc kubenswrapper[4646]: I1203 10:58:16.580474 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 10:58:16 crc kubenswrapper[4646]: I1203 10:58:16.598831 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 10:58:16 crc kubenswrapper[4646]: I1203 10:58:16.867690 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 10:58:16 crc kubenswrapper[4646]: I1203 10:58:16.885197 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 10:58:16 crc kubenswrapper[4646]: I1203 10:58:16.982009 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.038031 4646 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.050381 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.054978 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.080024 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.092940 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.227084 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.319116 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.364789 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.407383 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.422258 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.476278 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.518427 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.621478 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.633886 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.686786 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.770624 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.811122 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.843635 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.915965 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 10:58:17 crc kubenswrapper[4646]: I1203 10:58:17.953297 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.040673 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.054293 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.058870 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.164111 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.273291 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.313863 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.343757 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.350233 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.413328 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.442235 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.515892 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.598284 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.630650 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.665979 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.906410 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 10:58:18 crc kubenswrapper[4646]: I1203 10:58:18.956532 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.136087 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.137467 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.140164 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.172789 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.196283 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.226425 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.382463 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.414676 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.445274 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.546984 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.570791 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.668538 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.677438 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.865528 4646 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.930651 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.989325 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 10:58:19 crc kubenswrapper[4646]: I1203 10:58:19.992160 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.093191 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.097214 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.124103 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.146877 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.251469 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.274639 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.276096 4646 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.279835 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.279876 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.285230 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.306743 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=17.306730936 podStartE2EDuration="17.306730936s" podCreationTimestamp="2025-12-03 10:58:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:58:20.305946922 +0000 UTC m=+276.769003077" watchObservedRunningTime="2025-12-03 10:58:20.306730936 +0000 UTC m=+276.769787071" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.337996 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.404715 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.466545 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.508648 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.541018 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.579948 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.669534 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.710154 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.715869 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.781764 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 10:58:20 crc kubenswrapper[4646]: I1203 10:58:20.915137 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.037731 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.060315 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.111580 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.124765 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.134668 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.143704 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.244004 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.270535 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.325379 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.343621 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.374620 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.506891 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.522464 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.540459 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.576551 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.614034 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.618383 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.658669 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.704014 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.705114 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 10:58:21 crc kubenswrapper[4646]: I1203 10:58:21.729734 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.008931 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.009794 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.010697 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.011262 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.014613 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.014628 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.015538 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.015613 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.015857 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.025801 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.096186 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.129595 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.241858 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.260542 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.317000 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.532886 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.651232 4646 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.713753 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.729017 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.746798 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.874732 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 10:58:22 crc kubenswrapper[4646]: I1203 10:58:22.899399 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 10:58:23 crc kubenswrapper[4646]: I1203 10:58:23.016942 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 10:58:23 crc kubenswrapper[4646]: I1203 10:58:23.058439 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 10:58:23 crc kubenswrapper[4646]: I1203 10:58:23.194455 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 10:58:23 crc kubenswrapper[4646]: I1203 10:58:23.196603 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 10:58:23 crc kubenswrapper[4646]: I1203 10:58:23.227974 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 10:58:23 crc kubenswrapper[4646]: I1203 10:58:23.319648 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 10:58:23 crc kubenswrapper[4646]: I1203 10:58:23.322178 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 10:58:23 crc kubenswrapper[4646]: I1203 10:58:23.589988 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 10:58:23 crc kubenswrapper[4646]: I1203 10:58:23.842972 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 10:58:23 crc kubenswrapper[4646]: I1203 10:58:23.915736 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 10:58:23 crc kubenswrapper[4646]: I1203 10:58:23.929451 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 10:58:23 crc kubenswrapper[4646]: I1203 10:58:23.994971 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.030720 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.108703 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.136445 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.199183 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.219695 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.332093 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.339438 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.393855 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.536849 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.562412 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.609102 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.627710 4646 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.627935 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://625a623d805ff19721ca78009ff358b83659042674a5fae0abec2554449a54d5" gracePeriod=5 Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.678892 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.688264 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.745001 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.789594 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.863320 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.886059 4646 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 10:58:24 crc kubenswrapper[4646]: I1203 10:58:24.917047 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 10:58:25 crc kubenswrapper[4646]: I1203 10:58:25.057833 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 10:58:25 crc kubenswrapper[4646]: I1203 10:58:25.262022 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 10:58:25 crc kubenswrapper[4646]: I1203 10:58:25.333506 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 10:58:25 crc kubenswrapper[4646]: I1203 10:58:25.364707 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 10:58:25 crc kubenswrapper[4646]: I1203 10:58:25.585586 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 10:58:25 crc kubenswrapper[4646]: I1203 10:58:25.669578 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 10:58:25 crc kubenswrapper[4646]: I1203 10:58:25.734597 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 10:58:25 crc kubenswrapper[4646]: I1203 10:58:25.821708 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 10:58:25 crc kubenswrapper[4646]: I1203 10:58:25.852881 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 10:58:25 crc kubenswrapper[4646]: I1203 10:58:25.953891 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 10:58:25 crc kubenswrapper[4646]: I1203 10:58:25.997836 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.007104 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.048757 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.079051 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.140076 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.195913 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.217738 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.235882 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.309823 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.418421 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.590566 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.615779 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.638378 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.751345 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.783589 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.907993 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 10:58:26 crc kubenswrapper[4646]: I1203 10:58:26.959419 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.025668 4646 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.028100 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.083080 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.093622 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.234480 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.317388 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.360016 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.387941 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.464785 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.537962 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.725774 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.757236 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.776060 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.844533 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.908632 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 10:58:27 crc kubenswrapper[4646]: I1203 10:58:27.999261 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 10:58:28 crc kubenswrapper[4646]: I1203 10:58:28.038842 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 10:58:28 crc kubenswrapper[4646]: I1203 10:58:28.063229 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 10:58:28 crc kubenswrapper[4646]: I1203 10:58:28.084414 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 10:58:28 crc kubenswrapper[4646]: I1203 10:58:28.097500 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 10:58:28 crc kubenswrapper[4646]: I1203 10:58:28.147722 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 10:58:28 crc kubenswrapper[4646]: I1203 10:58:28.201261 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 10:58:28 crc kubenswrapper[4646]: I1203 10:58:28.213806 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 10:58:28 crc kubenswrapper[4646]: I1203 10:58:28.349438 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 10:58:28 crc kubenswrapper[4646]: I1203 10:58:28.392903 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 10:58:28 crc kubenswrapper[4646]: I1203 10:58:28.473586 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 10:58:28 crc kubenswrapper[4646]: I1203 10:58:28.501767 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 10:58:28 crc kubenswrapper[4646]: I1203 10:58:28.538197 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 10:58:28 crc kubenswrapper[4646]: I1203 10:58:28.653302 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 10:58:28 crc kubenswrapper[4646]: I1203 10:58:28.946901 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 10:58:29 crc kubenswrapper[4646]: I1203 10:58:29.261967 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 10:58:29 crc kubenswrapper[4646]: I1203 10:58:29.491853 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 10:58:29 crc kubenswrapper[4646]: I1203 10:58:29.743302 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 10:58:29 crc kubenswrapper[4646]: I1203 10:58:29.895279 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.192539 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.192649 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.196215 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.196270 4646 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="625a623d805ff19721ca78009ff358b83659042674a5fae0abec2554449a54d5" exitCode=137 Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.196355 4646 scope.go:117] "RemoveContainer" containerID="625a623d805ff19721ca78009ff358b83659042674a5fae0abec2554449a54d5" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.196495 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.215921 4646 scope.go:117] "RemoveContainer" containerID="625a623d805ff19721ca78009ff358b83659042674a5fae0abec2554449a54d5" Dec 03 10:58:30 crc kubenswrapper[4646]: E1203 10:58:30.216441 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"625a623d805ff19721ca78009ff358b83659042674a5fae0abec2554449a54d5\": container with ID starting with 625a623d805ff19721ca78009ff358b83659042674a5fae0abec2554449a54d5 not found: ID does not exist" containerID="625a623d805ff19721ca78009ff358b83659042674a5fae0abec2554449a54d5" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.216639 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"625a623d805ff19721ca78009ff358b83659042674a5fae0abec2554449a54d5"} err="failed to get container status \"625a623d805ff19721ca78009ff358b83659042674a5fae0abec2554449a54d5\": rpc error: code = NotFound desc = could not find container \"625a623d805ff19721ca78009ff358b83659042674a5fae0abec2554449a54d5\": container with ID starting with 625a623d805ff19721ca78009ff358b83659042674a5fae0abec2554449a54d5 not found: ID does not exist" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.365654 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.365761 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.365805 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.365852 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.365889 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.365913 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.365935 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.366016 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.366107 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.366364 4646 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.366390 4646 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.366409 4646 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.366428 4646 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.376771 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:58:30 crc kubenswrapper[4646]: I1203 10:58:30.468073 4646 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:31 crc kubenswrapper[4646]: I1203 10:58:31.855950 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.185175 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vkpc9"] Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.188934 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vkpc9" podUID="d1141f44-4444-47fb-878e-7a17db767b8d" containerName="registry-server" containerID="cri-o://659b6c320ce5f4246e99503f13f50a49dcf9ed082c911dd7b9d433cff30b3fd0" gracePeriod=30 Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.193638 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p9lsq"] Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.193919 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p9lsq" podUID="b13d8714-e047-4380-b095-7f5aac84f021" containerName="registry-server" containerID="cri-o://7d94c7c40f614008d49a7ee595dd5336eaafd156f1fc0cf9b92e0c291d1103f0" gracePeriod=30 Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.208174 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mbfc8"] Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.208782 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" podUID="143daef8-3687-4018-8f9f-2a675eac0909" containerName="marketplace-operator" containerID="cri-o://77faacb42b9bd8c4f04a3464f09b439de0a5af684fb7a1ce01ff4c6a6700e9ae" gracePeriod=30 Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.218394 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xjl9g"] Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.218995 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xjl9g" podUID="d20d9814-aa42-4109-9740-d0e5e6205a00" containerName="registry-server" containerID="cri-o://499dae3d8c1f749985ebe71582fe19a396f8a13081554ce6a9fec1916437e37f" gracePeriod=30 Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.221573 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vsmhw"] Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.221821 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vsmhw" podUID="250f331c-acf3-4b28-b2a4-515a8cceb844" containerName="registry-server" containerID="cri-o://d3af0c2b2856937cd01de17968d8a6e1f48a3e514fdd28700de3b42ce3aa34df" gracePeriod=30 Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.250549 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qvvn5"] Dec 03 10:58:38 crc kubenswrapper[4646]: E1203 10:58:38.250998 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9030be92-e97b-4760-a4cf-abe87c241ec6" containerName="installer" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.251083 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="9030be92-e97b-4760-a4cf-abe87c241ec6" containerName="installer" Dec 03 10:58:38 crc kubenswrapper[4646]: E1203 10:58:38.251161 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.251232 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.251448 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="9030be92-e97b-4760-a4cf-abe87c241ec6" containerName="installer" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.251546 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.252079 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.263715 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qvvn5"] Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.372775 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53441ce7-95f4-4236-8df4-fc04de470569-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qvvn5\" (UID: \"53441ce7-95f4-4236-8df4-fc04de470569\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.372854 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/53441ce7-95f4-4236-8df4-fc04de470569-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qvvn5\" (UID: \"53441ce7-95f4-4236-8df4-fc04de470569\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.373060 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lxqj\" (UniqueName: \"kubernetes.io/projected/53441ce7-95f4-4236-8df4-fc04de470569-kube-api-access-4lxqj\") pod \"marketplace-operator-79b997595-qvvn5\" (UID: \"53441ce7-95f4-4236-8df4-fc04de470569\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.474063 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lxqj\" (UniqueName: \"kubernetes.io/projected/53441ce7-95f4-4236-8df4-fc04de470569-kube-api-access-4lxqj\") pod \"marketplace-operator-79b997595-qvvn5\" (UID: \"53441ce7-95f4-4236-8df4-fc04de470569\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.474115 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53441ce7-95f4-4236-8df4-fc04de470569-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qvvn5\" (UID: \"53441ce7-95f4-4236-8df4-fc04de470569\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.474144 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/53441ce7-95f4-4236-8df4-fc04de470569-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qvvn5\" (UID: \"53441ce7-95f4-4236-8df4-fc04de470569\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.476103 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53441ce7-95f4-4236-8df4-fc04de470569-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qvvn5\" (UID: \"53441ce7-95f4-4236-8df4-fc04de470569\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.482022 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/53441ce7-95f4-4236-8df4-fc04de470569-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qvvn5\" (UID: \"53441ce7-95f4-4236-8df4-fc04de470569\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.490068 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lxqj\" (UniqueName: \"kubernetes.io/projected/53441ce7-95f4-4236-8df4-fc04de470569-kube-api-access-4lxqj\") pod \"marketplace-operator-79b997595-qvvn5\" (UID: \"53441ce7-95f4-4236-8df4-fc04de470569\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.560006 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.560095 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.589642 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.611185 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.613024 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.653983 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.678042 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ts5nb\" (UniqueName: \"kubernetes.io/projected/d1141f44-4444-47fb-878e-7a17db767b8d-kube-api-access-ts5nb\") pod \"d1141f44-4444-47fb-878e-7a17db767b8d\" (UID: \"d1141f44-4444-47fb-878e-7a17db767b8d\") " Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.678078 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1141f44-4444-47fb-878e-7a17db767b8d-utilities\") pod \"d1141f44-4444-47fb-878e-7a17db767b8d\" (UID: \"d1141f44-4444-47fb-878e-7a17db767b8d\") " Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.678101 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1141f44-4444-47fb-878e-7a17db767b8d-catalog-content\") pod \"d1141f44-4444-47fb-878e-7a17db767b8d\" (UID: \"d1141f44-4444-47fb-878e-7a17db767b8d\") " Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.678162 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b13d8714-e047-4380-b095-7f5aac84f021-catalog-content\") pod \"b13d8714-e047-4380-b095-7f5aac84f021\" (UID: \"b13d8714-e047-4380-b095-7f5aac84f021\") " Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.678181 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/143daef8-3687-4018-8f9f-2a675eac0909-marketplace-trusted-ca\") pod \"143daef8-3687-4018-8f9f-2a675eac0909\" (UID: \"143daef8-3687-4018-8f9f-2a675eac0909\") " Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.678203 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/143daef8-3687-4018-8f9f-2a675eac0909-marketplace-operator-metrics\") pod \"143daef8-3687-4018-8f9f-2a675eac0909\" (UID: \"143daef8-3687-4018-8f9f-2a675eac0909\") " Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.678222 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g62tw\" (UniqueName: \"kubernetes.io/projected/b13d8714-e047-4380-b095-7f5aac84f021-kube-api-access-g62tw\") pod \"b13d8714-e047-4380-b095-7f5aac84f021\" (UID: \"b13d8714-e047-4380-b095-7f5aac84f021\") " Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.678246 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d20d9814-aa42-4109-9740-d0e5e6205a00-utilities\") pod \"d20d9814-aa42-4109-9740-d0e5e6205a00\" (UID: \"d20d9814-aa42-4109-9740-d0e5e6205a00\") " Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.678270 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghzsx\" (UniqueName: \"kubernetes.io/projected/143daef8-3687-4018-8f9f-2a675eac0909-kube-api-access-ghzsx\") pod \"143daef8-3687-4018-8f9f-2a675eac0909\" (UID: \"143daef8-3687-4018-8f9f-2a675eac0909\") " Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.678296 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxd62\" (UniqueName: \"kubernetes.io/projected/d20d9814-aa42-4109-9740-d0e5e6205a00-kube-api-access-xxd62\") pod \"d20d9814-aa42-4109-9740-d0e5e6205a00\" (UID: \"d20d9814-aa42-4109-9740-d0e5e6205a00\") " Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.678313 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d20d9814-aa42-4109-9740-d0e5e6205a00-catalog-content\") pod \"d20d9814-aa42-4109-9740-d0e5e6205a00\" (UID: \"d20d9814-aa42-4109-9740-d0e5e6205a00\") " Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.678930 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b13d8714-e047-4380-b095-7f5aac84f021-utilities\") pod \"b13d8714-e047-4380-b095-7f5aac84f021\" (UID: \"b13d8714-e047-4380-b095-7f5aac84f021\") " Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.679136 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1141f44-4444-47fb-878e-7a17db767b8d-utilities" (OuterVolumeSpecName: "utilities") pod "d1141f44-4444-47fb-878e-7a17db767b8d" (UID: "d1141f44-4444-47fb-878e-7a17db767b8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.684254 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1141f44-4444-47fb-878e-7a17db767b8d-kube-api-access-ts5nb" (OuterVolumeSpecName: "kube-api-access-ts5nb") pod "d1141f44-4444-47fb-878e-7a17db767b8d" (UID: "d1141f44-4444-47fb-878e-7a17db767b8d"). InnerVolumeSpecName "kube-api-access-ts5nb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.686164 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b13d8714-e047-4380-b095-7f5aac84f021-utilities" (OuterVolumeSpecName: "utilities") pod "b13d8714-e047-4380-b095-7f5aac84f021" (UID: "b13d8714-e047-4380-b095-7f5aac84f021"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.687628 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d20d9814-aa42-4109-9740-d0e5e6205a00-utilities" (OuterVolumeSpecName: "utilities") pod "d20d9814-aa42-4109-9740-d0e5e6205a00" (UID: "d20d9814-aa42-4109-9740-d0e5e6205a00"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.688747 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/143daef8-3687-4018-8f9f-2a675eac0909-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "143daef8-3687-4018-8f9f-2a675eac0909" (UID: "143daef8-3687-4018-8f9f-2a675eac0909"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.689862 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b13d8714-e047-4380-b095-7f5aac84f021-kube-api-access-g62tw" (OuterVolumeSpecName: "kube-api-access-g62tw") pod "b13d8714-e047-4380-b095-7f5aac84f021" (UID: "b13d8714-e047-4380-b095-7f5aac84f021"). InnerVolumeSpecName "kube-api-access-g62tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.690071 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d20d9814-aa42-4109-9740-d0e5e6205a00-kube-api-access-xxd62" (OuterVolumeSpecName: "kube-api-access-xxd62") pod "d20d9814-aa42-4109-9740-d0e5e6205a00" (UID: "d20d9814-aa42-4109-9740-d0e5e6205a00"). InnerVolumeSpecName "kube-api-access-xxd62". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.691897 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/143daef8-3687-4018-8f9f-2a675eac0909-kube-api-access-ghzsx" (OuterVolumeSpecName: "kube-api-access-ghzsx") pod "143daef8-3687-4018-8f9f-2a675eac0909" (UID: "143daef8-3687-4018-8f9f-2a675eac0909"). InnerVolumeSpecName "kube-api-access-ghzsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.694951 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/143daef8-3687-4018-8f9f-2a675eac0909-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "143daef8-3687-4018-8f9f-2a675eac0909" (UID: "143daef8-3687-4018-8f9f-2a675eac0909"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.727318 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d20d9814-aa42-4109-9740-d0e5e6205a00-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d20d9814-aa42-4109-9740-d0e5e6205a00" (UID: "d20d9814-aa42-4109-9740-d0e5e6205a00"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.752491 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1141f44-4444-47fb-878e-7a17db767b8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1141f44-4444-47fb-878e-7a17db767b8d" (UID: "d1141f44-4444-47fb-878e-7a17db767b8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.767417 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b13d8714-e047-4380-b095-7f5aac84f021-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b13d8714-e047-4380-b095-7f5aac84f021" (UID: "b13d8714-e047-4380-b095-7f5aac84f021"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.780399 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/250f331c-acf3-4b28-b2a4-515a8cceb844-utilities\") pod \"250f331c-acf3-4b28-b2a4-515a8cceb844\" (UID: \"250f331c-acf3-4b28-b2a4-515a8cceb844\") " Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.780541 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6vrt\" (UniqueName: \"kubernetes.io/projected/250f331c-acf3-4b28-b2a4-515a8cceb844-kube-api-access-m6vrt\") pod \"250f331c-acf3-4b28-b2a4-515a8cceb844\" (UID: \"250f331c-acf3-4b28-b2a4-515a8cceb844\") " Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.780576 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/250f331c-acf3-4b28-b2a4-515a8cceb844-catalog-content\") pod \"250f331c-acf3-4b28-b2a4-515a8cceb844\" (UID: \"250f331c-acf3-4b28-b2a4-515a8cceb844\") " Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.780870 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b13d8714-e047-4380-b095-7f5aac84f021-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.780890 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ts5nb\" (UniqueName: \"kubernetes.io/projected/d1141f44-4444-47fb-878e-7a17db767b8d-kube-api-access-ts5nb\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.780901 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1141f44-4444-47fb-878e-7a17db767b8d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.780914 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1141f44-4444-47fb-878e-7a17db767b8d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.780924 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b13d8714-e047-4380-b095-7f5aac84f021-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.780935 4646 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/143daef8-3687-4018-8f9f-2a675eac0909-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.780947 4646 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/143daef8-3687-4018-8f9f-2a675eac0909-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.780959 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g62tw\" (UniqueName: \"kubernetes.io/projected/b13d8714-e047-4380-b095-7f5aac84f021-kube-api-access-g62tw\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.780970 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d20d9814-aa42-4109-9740-d0e5e6205a00-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.780981 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghzsx\" (UniqueName: \"kubernetes.io/projected/143daef8-3687-4018-8f9f-2a675eac0909-kube-api-access-ghzsx\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.780992 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d20d9814-aa42-4109-9740-d0e5e6205a00-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.781002 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxd62\" (UniqueName: \"kubernetes.io/projected/d20d9814-aa42-4109-9740-d0e5e6205a00-kube-api-access-xxd62\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.784101 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/250f331c-acf3-4b28-b2a4-515a8cceb844-kube-api-access-m6vrt" (OuterVolumeSpecName: "kube-api-access-m6vrt") pod "250f331c-acf3-4b28-b2a4-515a8cceb844" (UID: "250f331c-acf3-4b28-b2a4-515a8cceb844"). InnerVolumeSpecName "kube-api-access-m6vrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.784884 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/250f331c-acf3-4b28-b2a4-515a8cceb844-utilities" (OuterVolumeSpecName: "utilities") pod "250f331c-acf3-4b28-b2a4-515a8cceb844" (UID: "250f331c-acf3-4b28-b2a4-515a8cceb844"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.786301 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qvvn5"] Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.881933 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6vrt\" (UniqueName: \"kubernetes.io/projected/250f331c-acf3-4b28-b2a4-515a8cceb844-kube-api-access-m6vrt\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.881970 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/250f331c-acf3-4b28-b2a4-515a8cceb844-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.897019 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/250f331c-acf3-4b28-b2a4-515a8cceb844-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "250f331c-acf3-4b28-b2a4-515a8cceb844" (UID: "250f331c-acf3-4b28-b2a4-515a8cceb844"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:58:38 crc kubenswrapper[4646]: I1203 10:58:38.983784 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/250f331c-acf3-4b28-b2a4-515a8cceb844-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.249471 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" event={"ID":"53441ce7-95f4-4236-8df4-fc04de470569","Type":"ContainerStarted","Data":"3b9684873f6ae440e4e3b50ef9fc44d8013a78d54e64bf279ea39174eba3fb9d"} Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.249530 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" event={"ID":"53441ce7-95f4-4236-8df4-fc04de470569","Type":"ContainerStarted","Data":"4064bbcf2bef8634f481a598809df9c96ef1c4a5b844a5d04adfee0bcb6ea051"} Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.251092 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.252101 4646 generic.go:334] "Generic (PLEG): container finished" podID="d20d9814-aa42-4109-9740-d0e5e6205a00" containerID="499dae3d8c1f749985ebe71582fe19a396f8a13081554ce6a9fec1916437e37f" exitCode=0 Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.252173 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xjl9g" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.252183 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjl9g" event={"ID":"d20d9814-aa42-4109-9740-d0e5e6205a00","Type":"ContainerDied","Data":"499dae3d8c1f749985ebe71582fe19a396f8a13081554ce6a9fec1916437e37f"} Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.252467 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.252561 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjl9g" event={"ID":"d20d9814-aa42-4109-9740-d0e5e6205a00","Type":"ContainerDied","Data":"e9c7a8c9b8cf5a21833988a2e3ad4a8fcf3cae119cdf014cf47488f75b97e880"} Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.252499 4646 scope.go:117] "RemoveContainer" containerID="499dae3d8c1f749985ebe71582fe19a396f8a13081554ce6a9fec1916437e37f" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.256170 4646 generic.go:334] "Generic (PLEG): container finished" podID="d1141f44-4444-47fb-878e-7a17db767b8d" containerID="659b6c320ce5f4246e99503f13f50a49dcf9ed082c911dd7b9d433cff30b3fd0" exitCode=0 Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.256201 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkpc9" event={"ID":"d1141f44-4444-47fb-878e-7a17db767b8d","Type":"ContainerDied","Data":"659b6c320ce5f4246e99503f13f50a49dcf9ed082c911dd7b9d433cff30b3fd0"} Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.256229 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkpc9" event={"ID":"d1141f44-4444-47fb-878e-7a17db767b8d","Type":"ContainerDied","Data":"fd0584fef9cbb97cae3b97f8c6905ac586fb2de2cfd684fa21c8a5babd791f44"} Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.256240 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkpc9" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.258054 4646 generic.go:334] "Generic (PLEG): container finished" podID="250f331c-acf3-4b28-b2a4-515a8cceb844" containerID="d3af0c2b2856937cd01de17968d8a6e1f48a3e514fdd28700de3b42ce3aa34df" exitCode=0 Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.258100 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsmhw" event={"ID":"250f331c-acf3-4b28-b2a4-515a8cceb844","Type":"ContainerDied","Data":"d3af0c2b2856937cd01de17968d8a6e1f48a3e514fdd28700de3b42ce3aa34df"} Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.258114 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsmhw" event={"ID":"250f331c-acf3-4b28-b2a4-515a8cceb844","Type":"ContainerDied","Data":"0eaf5f3c287678ceb54cf44f2f32c7ad804002b076b094183b514a348f31beeb"} Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.258173 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vsmhw" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.271692 4646 generic.go:334] "Generic (PLEG): container finished" podID="143daef8-3687-4018-8f9f-2a675eac0909" containerID="77faacb42b9bd8c4f04a3464f09b439de0a5af684fb7a1ce01ff4c6a6700e9ae" exitCode=0 Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.271824 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" event={"ID":"143daef8-3687-4018-8f9f-2a675eac0909","Type":"ContainerDied","Data":"77faacb42b9bd8c4f04a3464f09b439de0a5af684fb7a1ce01ff4c6a6700e9ae"} Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.271858 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" event={"ID":"143daef8-3687-4018-8f9f-2a675eac0909","Type":"ContainerDied","Data":"f9b27d11775849302b28543ffe31fe8721e2ab8538efaa59f7c8c40be1ae7e33"} Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.272091 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mbfc8" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.274985 4646 generic.go:334] "Generic (PLEG): container finished" podID="b13d8714-e047-4380-b095-7f5aac84f021" containerID="7d94c7c40f614008d49a7ee595dd5336eaafd156f1fc0cf9b92e0c291d1103f0" exitCode=0 Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.275012 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p9lsq" event={"ID":"b13d8714-e047-4380-b095-7f5aac84f021","Type":"ContainerDied","Data":"7d94c7c40f614008d49a7ee595dd5336eaafd156f1fc0cf9b92e0c291d1103f0"} Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.275026 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p9lsq" event={"ID":"b13d8714-e047-4380-b095-7f5aac84f021","Type":"ContainerDied","Data":"7f71624badfa9a269a1e00b1617569675cad4d6c15b7a3bef791825b406ea716"} Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.275071 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p9lsq" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.279427 4646 scope.go:117] "RemoveContainer" containerID="14b48e0f62154c7f2c9b9a3758864283bee56cea044cc60ede045e4395aa6cdb" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.311098 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-qvvn5" podStartSLOduration=1.311077437 podStartE2EDuration="1.311077437s" podCreationTimestamp="2025-12-03 10:58:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:58:39.279669689 +0000 UTC m=+295.742725844" watchObservedRunningTime="2025-12-03 10:58:39.311077437 +0000 UTC m=+295.774133582" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.311665 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xjl9g"] Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.315695 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xjl9g"] Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.334364 4646 scope.go:117] "RemoveContainer" containerID="d82a373f45467842e1c0037753d5e6c184d58e75507db49e79657d1bd0052da4" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.355051 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vkpc9"] Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.360931 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vkpc9"] Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.377487 4646 scope.go:117] "RemoveContainer" containerID="499dae3d8c1f749985ebe71582fe19a396f8a13081554ce6a9fec1916437e37f" Dec 03 10:58:39 crc kubenswrapper[4646]: E1203 10:58:39.378798 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"499dae3d8c1f749985ebe71582fe19a396f8a13081554ce6a9fec1916437e37f\": container with ID starting with 499dae3d8c1f749985ebe71582fe19a396f8a13081554ce6a9fec1916437e37f not found: ID does not exist" containerID="499dae3d8c1f749985ebe71582fe19a396f8a13081554ce6a9fec1916437e37f" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.378829 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"499dae3d8c1f749985ebe71582fe19a396f8a13081554ce6a9fec1916437e37f"} err="failed to get container status \"499dae3d8c1f749985ebe71582fe19a396f8a13081554ce6a9fec1916437e37f\": rpc error: code = NotFound desc = could not find container \"499dae3d8c1f749985ebe71582fe19a396f8a13081554ce6a9fec1916437e37f\": container with ID starting with 499dae3d8c1f749985ebe71582fe19a396f8a13081554ce6a9fec1916437e37f not found: ID does not exist" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.379117 4646 scope.go:117] "RemoveContainer" containerID="14b48e0f62154c7f2c9b9a3758864283bee56cea044cc60ede045e4395aa6cdb" Dec 03 10:58:39 crc kubenswrapper[4646]: E1203 10:58:39.379627 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14b48e0f62154c7f2c9b9a3758864283bee56cea044cc60ede045e4395aa6cdb\": container with ID starting with 14b48e0f62154c7f2c9b9a3758864283bee56cea044cc60ede045e4395aa6cdb not found: ID does not exist" containerID="14b48e0f62154c7f2c9b9a3758864283bee56cea044cc60ede045e4395aa6cdb" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.379656 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14b48e0f62154c7f2c9b9a3758864283bee56cea044cc60ede045e4395aa6cdb"} err="failed to get container status \"14b48e0f62154c7f2c9b9a3758864283bee56cea044cc60ede045e4395aa6cdb\": rpc error: code = NotFound desc = could not find container \"14b48e0f62154c7f2c9b9a3758864283bee56cea044cc60ede045e4395aa6cdb\": container with ID starting with 14b48e0f62154c7f2c9b9a3758864283bee56cea044cc60ede045e4395aa6cdb not found: ID does not exist" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.379671 4646 scope.go:117] "RemoveContainer" containerID="d82a373f45467842e1c0037753d5e6c184d58e75507db49e79657d1bd0052da4" Dec 03 10:58:39 crc kubenswrapper[4646]: E1203 10:58:39.380272 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d82a373f45467842e1c0037753d5e6c184d58e75507db49e79657d1bd0052da4\": container with ID starting with d82a373f45467842e1c0037753d5e6c184d58e75507db49e79657d1bd0052da4 not found: ID does not exist" containerID="d82a373f45467842e1c0037753d5e6c184d58e75507db49e79657d1bd0052da4" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.380291 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d82a373f45467842e1c0037753d5e6c184d58e75507db49e79657d1bd0052da4"} err="failed to get container status \"d82a373f45467842e1c0037753d5e6c184d58e75507db49e79657d1bd0052da4\": rpc error: code = NotFound desc = could not find container \"d82a373f45467842e1c0037753d5e6c184d58e75507db49e79657d1bd0052da4\": container with ID starting with d82a373f45467842e1c0037753d5e6c184d58e75507db49e79657d1bd0052da4 not found: ID does not exist" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.380305 4646 scope.go:117] "RemoveContainer" containerID="659b6c320ce5f4246e99503f13f50a49dcf9ed082c911dd7b9d433cff30b3fd0" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.383277 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p9lsq"] Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.389221 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p9lsq"] Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.406921 4646 scope.go:117] "RemoveContainer" containerID="7f97b5c971fc00c9628d805dd112cffaee6de0097a01311b4b0cdb7f949b5085" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.407814 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vsmhw"] Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.410809 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vsmhw"] Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.413448 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mbfc8"] Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.416037 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mbfc8"] Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.420400 4646 scope.go:117] "RemoveContainer" containerID="f69d040041dfd1c522638e466cf336112bf5c7f43b79f49a4e6160f2ddf229df" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.434503 4646 scope.go:117] "RemoveContainer" containerID="659b6c320ce5f4246e99503f13f50a49dcf9ed082c911dd7b9d433cff30b3fd0" Dec 03 10:58:39 crc kubenswrapper[4646]: E1203 10:58:39.435031 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"659b6c320ce5f4246e99503f13f50a49dcf9ed082c911dd7b9d433cff30b3fd0\": container with ID starting with 659b6c320ce5f4246e99503f13f50a49dcf9ed082c911dd7b9d433cff30b3fd0 not found: ID does not exist" containerID="659b6c320ce5f4246e99503f13f50a49dcf9ed082c911dd7b9d433cff30b3fd0" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.435081 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"659b6c320ce5f4246e99503f13f50a49dcf9ed082c911dd7b9d433cff30b3fd0"} err="failed to get container status \"659b6c320ce5f4246e99503f13f50a49dcf9ed082c911dd7b9d433cff30b3fd0\": rpc error: code = NotFound desc = could not find container \"659b6c320ce5f4246e99503f13f50a49dcf9ed082c911dd7b9d433cff30b3fd0\": container with ID starting with 659b6c320ce5f4246e99503f13f50a49dcf9ed082c911dd7b9d433cff30b3fd0 not found: ID does not exist" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.435105 4646 scope.go:117] "RemoveContainer" containerID="7f97b5c971fc00c9628d805dd112cffaee6de0097a01311b4b0cdb7f949b5085" Dec 03 10:58:39 crc kubenswrapper[4646]: E1203 10:58:39.435417 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f97b5c971fc00c9628d805dd112cffaee6de0097a01311b4b0cdb7f949b5085\": container with ID starting with 7f97b5c971fc00c9628d805dd112cffaee6de0097a01311b4b0cdb7f949b5085 not found: ID does not exist" containerID="7f97b5c971fc00c9628d805dd112cffaee6de0097a01311b4b0cdb7f949b5085" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.435438 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f97b5c971fc00c9628d805dd112cffaee6de0097a01311b4b0cdb7f949b5085"} err="failed to get container status \"7f97b5c971fc00c9628d805dd112cffaee6de0097a01311b4b0cdb7f949b5085\": rpc error: code = NotFound desc = could not find container \"7f97b5c971fc00c9628d805dd112cffaee6de0097a01311b4b0cdb7f949b5085\": container with ID starting with 7f97b5c971fc00c9628d805dd112cffaee6de0097a01311b4b0cdb7f949b5085 not found: ID does not exist" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.435454 4646 scope.go:117] "RemoveContainer" containerID="f69d040041dfd1c522638e466cf336112bf5c7f43b79f49a4e6160f2ddf229df" Dec 03 10:58:39 crc kubenswrapper[4646]: E1203 10:58:39.435792 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f69d040041dfd1c522638e466cf336112bf5c7f43b79f49a4e6160f2ddf229df\": container with ID starting with f69d040041dfd1c522638e466cf336112bf5c7f43b79f49a4e6160f2ddf229df not found: ID does not exist" containerID="f69d040041dfd1c522638e466cf336112bf5c7f43b79f49a4e6160f2ddf229df" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.435814 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f69d040041dfd1c522638e466cf336112bf5c7f43b79f49a4e6160f2ddf229df"} err="failed to get container status \"f69d040041dfd1c522638e466cf336112bf5c7f43b79f49a4e6160f2ddf229df\": rpc error: code = NotFound desc = could not find container \"f69d040041dfd1c522638e466cf336112bf5c7f43b79f49a4e6160f2ddf229df\": container with ID starting with f69d040041dfd1c522638e466cf336112bf5c7f43b79f49a4e6160f2ddf229df not found: ID does not exist" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.435828 4646 scope.go:117] "RemoveContainer" containerID="d3af0c2b2856937cd01de17968d8a6e1f48a3e514fdd28700de3b42ce3aa34df" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.449174 4646 scope.go:117] "RemoveContainer" containerID="b48c86e7cc849df6c509ba02fd6c836621c1d2cf0e125e46861115af2795ad36" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.464558 4646 scope.go:117] "RemoveContainer" containerID="a75b773fc6a2e0df7b0b6ab650b485a193bc7a61fa8320aae2b8c005e1c7dcc1" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.480768 4646 scope.go:117] "RemoveContainer" containerID="d3af0c2b2856937cd01de17968d8a6e1f48a3e514fdd28700de3b42ce3aa34df" Dec 03 10:58:39 crc kubenswrapper[4646]: E1203 10:58:39.481541 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3af0c2b2856937cd01de17968d8a6e1f48a3e514fdd28700de3b42ce3aa34df\": container with ID starting with d3af0c2b2856937cd01de17968d8a6e1f48a3e514fdd28700de3b42ce3aa34df not found: ID does not exist" containerID="d3af0c2b2856937cd01de17968d8a6e1f48a3e514fdd28700de3b42ce3aa34df" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.481591 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3af0c2b2856937cd01de17968d8a6e1f48a3e514fdd28700de3b42ce3aa34df"} err="failed to get container status \"d3af0c2b2856937cd01de17968d8a6e1f48a3e514fdd28700de3b42ce3aa34df\": rpc error: code = NotFound desc = could not find container \"d3af0c2b2856937cd01de17968d8a6e1f48a3e514fdd28700de3b42ce3aa34df\": container with ID starting with d3af0c2b2856937cd01de17968d8a6e1f48a3e514fdd28700de3b42ce3aa34df not found: ID does not exist" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.481648 4646 scope.go:117] "RemoveContainer" containerID="b48c86e7cc849df6c509ba02fd6c836621c1d2cf0e125e46861115af2795ad36" Dec 03 10:58:39 crc kubenswrapper[4646]: E1203 10:58:39.482013 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b48c86e7cc849df6c509ba02fd6c836621c1d2cf0e125e46861115af2795ad36\": container with ID starting with b48c86e7cc849df6c509ba02fd6c836621c1d2cf0e125e46861115af2795ad36 not found: ID does not exist" containerID="b48c86e7cc849df6c509ba02fd6c836621c1d2cf0e125e46861115af2795ad36" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.482042 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b48c86e7cc849df6c509ba02fd6c836621c1d2cf0e125e46861115af2795ad36"} err="failed to get container status \"b48c86e7cc849df6c509ba02fd6c836621c1d2cf0e125e46861115af2795ad36\": rpc error: code = NotFound desc = could not find container \"b48c86e7cc849df6c509ba02fd6c836621c1d2cf0e125e46861115af2795ad36\": container with ID starting with b48c86e7cc849df6c509ba02fd6c836621c1d2cf0e125e46861115af2795ad36 not found: ID does not exist" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.482066 4646 scope.go:117] "RemoveContainer" containerID="a75b773fc6a2e0df7b0b6ab650b485a193bc7a61fa8320aae2b8c005e1c7dcc1" Dec 03 10:58:39 crc kubenswrapper[4646]: E1203 10:58:39.482326 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a75b773fc6a2e0df7b0b6ab650b485a193bc7a61fa8320aae2b8c005e1c7dcc1\": container with ID starting with a75b773fc6a2e0df7b0b6ab650b485a193bc7a61fa8320aae2b8c005e1c7dcc1 not found: ID does not exist" containerID="a75b773fc6a2e0df7b0b6ab650b485a193bc7a61fa8320aae2b8c005e1c7dcc1" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.482398 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a75b773fc6a2e0df7b0b6ab650b485a193bc7a61fa8320aae2b8c005e1c7dcc1"} err="failed to get container status \"a75b773fc6a2e0df7b0b6ab650b485a193bc7a61fa8320aae2b8c005e1c7dcc1\": rpc error: code = NotFound desc = could not find container \"a75b773fc6a2e0df7b0b6ab650b485a193bc7a61fa8320aae2b8c005e1c7dcc1\": container with ID starting with a75b773fc6a2e0df7b0b6ab650b485a193bc7a61fa8320aae2b8c005e1c7dcc1 not found: ID does not exist" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.482414 4646 scope.go:117] "RemoveContainer" containerID="77faacb42b9bd8c4f04a3464f09b439de0a5af684fb7a1ce01ff4c6a6700e9ae" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.493933 4646 scope.go:117] "RemoveContainer" containerID="77faacb42b9bd8c4f04a3464f09b439de0a5af684fb7a1ce01ff4c6a6700e9ae" Dec 03 10:58:39 crc kubenswrapper[4646]: E1203 10:58:39.494170 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77faacb42b9bd8c4f04a3464f09b439de0a5af684fb7a1ce01ff4c6a6700e9ae\": container with ID starting with 77faacb42b9bd8c4f04a3464f09b439de0a5af684fb7a1ce01ff4c6a6700e9ae not found: ID does not exist" containerID="77faacb42b9bd8c4f04a3464f09b439de0a5af684fb7a1ce01ff4c6a6700e9ae" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.494193 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77faacb42b9bd8c4f04a3464f09b439de0a5af684fb7a1ce01ff4c6a6700e9ae"} err="failed to get container status \"77faacb42b9bd8c4f04a3464f09b439de0a5af684fb7a1ce01ff4c6a6700e9ae\": rpc error: code = NotFound desc = could not find container \"77faacb42b9bd8c4f04a3464f09b439de0a5af684fb7a1ce01ff4c6a6700e9ae\": container with ID starting with 77faacb42b9bd8c4f04a3464f09b439de0a5af684fb7a1ce01ff4c6a6700e9ae not found: ID does not exist" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.494212 4646 scope.go:117] "RemoveContainer" containerID="7d94c7c40f614008d49a7ee595dd5336eaafd156f1fc0cf9b92e0c291d1103f0" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.505552 4646 scope.go:117] "RemoveContainer" containerID="d83e90241c43fa5c9ec4c4dc28ad273aadf58f89b0ad1edcc0909d092a56b2a2" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.518733 4646 scope.go:117] "RemoveContainer" containerID="31326048919f5e8ecb016a0b00ce9fc87aa78a5c06ac18e588412e93940a0249" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.533451 4646 scope.go:117] "RemoveContainer" containerID="7d94c7c40f614008d49a7ee595dd5336eaafd156f1fc0cf9b92e0c291d1103f0" Dec 03 10:58:39 crc kubenswrapper[4646]: E1203 10:58:39.533806 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d94c7c40f614008d49a7ee595dd5336eaafd156f1fc0cf9b92e0c291d1103f0\": container with ID starting with 7d94c7c40f614008d49a7ee595dd5336eaafd156f1fc0cf9b92e0c291d1103f0 not found: ID does not exist" containerID="7d94c7c40f614008d49a7ee595dd5336eaafd156f1fc0cf9b92e0c291d1103f0" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.533832 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d94c7c40f614008d49a7ee595dd5336eaafd156f1fc0cf9b92e0c291d1103f0"} err="failed to get container status \"7d94c7c40f614008d49a7ee595dd5336eaafd156f1fc0cf9b92e0c291d1103f0\": rpc error: code = NotFound desc = could not find container \"7d94c7c40f614008d49a7ee595dd5336eaafd156f1fc0cf9b92e0c291d1103f0\": container with ID starting with 7d94c7c40f614008d49a7ee595dd5336eaafd156f1fc0cf9b92e0c291d1103f0 not found: ID does not exist" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.533869 4646 scope.go:117] "RemoveContainer" containerID="d83e90241c43fa5c9ec4c4dc28ad273aadf58f89b0ad1edcc0909d092a56b2a2" Dec 03 10:58:39 crc kubenswrapper[4646]: E1203 10:58:39.534160 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d83e90241c43fa5c9ec4c4dc28ad273aadf58f89b0ad1edcc0909d092a56b2a2\": container with ID starting with d83e90241c43fa5c9ec4c4dc28ad273aadf58f89b0ad1edcc0909d092a56b2a2 not found: ID does not exist" containerID="d83e90241c43fa5c9ec4c4dc28ad273aadf58f89b0ad1edcc0909d092a56b2a2" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.534179 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d83e90241c43fa5c9ec4c4dc28ad273aadf58f89b0ad1edcc0909d092a56b2a2"} err="failed to get container status \"d83e90241c43fa5c9ec4c4dc28ad273aadf58f89b0ad1edcc0909d092a56b2a2\": rpc error: code = NotFound desc = could not find container \"d83e90241c43fa5c9ec4c4dc28ad273aadf58f89b0ad1edcc0909d092a56b2a2\": container with ID starting with d83e90241c43fa5c9ec4c4dc28ad273aadf58f89b0ad1edcc0909d092a56b2a2 not found: ID does not exist" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.534208 4646 scope.go:117] "RemoveContainer" containerID="31326048919f5e8ecb016a0b00ce9fc87aa78a5c06ac18e588412e93940a0249" Dec 03 10:58:39 crc kubenswrapper[4646]: E1203 10:58:39.534496 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31326048919f5e8ecb016a0b00ce9fc87aa78a5c06ac18e588412e93940a0249\": container with ID starting with 31326048919f5e8ecb016a0b00ce9fc87aa78a5c06ac18e588412e93940a0249 not found: ID does not exist" containerID="31326048919f5e8ecb016a0b00ce9fc87aa78a5c06ac18e588412e93940a0249" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.534516 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31326048919f5e8ecb016a0b00ce9fc87aa78a5c06ac18e588412e93940a0249"} err="failed to get container status \"31326048919f5e8ecb016a0b00ce9fc87aa78a5c06ac18e588412e93940a0249\": rpc error: code = NotFound desc = could not find container \"31326048919f5e8ecb016a0b00ce9fc87aa78a5c06ac18e588412e93940a0249\": container with ID starting with 31326048919f5e8ecb016a0b00ce9fc87aa78a5c06ac18e588412e93940a0249 not found: ID does not exist" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.860137 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="143daef8-3687-4018-8f9f-2a675eac0909" path="/var/lib/kubelet/pods/143daef8-3687-4018-8f9f-2a675eac0909/volumes" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.861575 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="250f331c-acf3-4b28-b2a4-515a8cceb844" path="/var/lib/kubelet/pods/250f331c-acf3-4b28-b2a4-515a8cceb844/volumes" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.862962 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b13d8714-e047-4380-b095-7f5aac84f021" path="/var/lib/kubelet/pods/b13d8714-e047-4380-b095-7f5aac84f021/volumes" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.865274 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1141f44-4444-47fb-878e-7a17db767b8d" path="/var/lib/kubelet/pods/d1141f44-4444-47fb-878e-7a17db767b8d/volumes" Dec 03 10:58:39 crc kubenswrapper[4646]: I1203 10:58:39.866982 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d20d9814-aa42-4109-9740-d0e5e6205a00" path="/var/lib/kubelet/pods/d20d9814-aa42-4109-9740-d0e5e6205a00/volumes" Dec 03 10:58:46 crc kubenswrapper[4646]: I1203 10:58:46.025758 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 10:58:50 crc kubenswrapper[4646]: I1203 10:58:50.833663 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 10:58:50 crc kubenswrapper[4646]: I1203 10:58:50.933409 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 10:58:53 crc kubenswrapper[4646]: I1203 10:58:53.742069 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 10:58:57 crc kubenswrapper[4646]: I1203 10:58:57.568638 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 10:58:57 crc kubenswrapper[4646]: I1203 10:58:57.629371 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 10:58:57 crc kubenswrapper[4646]: I1203 10:58:57.792441 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-czl94"] Dec 03 10:58:57 crc kubenswrapper[4646]: I1203 10:58:57.792652 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" podUID="e51eada8-802d-4064-8705-2647ffa44189" containerName="controller-manager" containerID="cri-o://f5b280f3f57a36058f5aa0b2f298d810688b521848dda8b6ce675fe80b058e23" gracePeriod=30 Dec 03 10:58:57 crc kubenswrapper[4646]: I1203 10:58:57.896205 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj"] Dec 03 10:58:57 crc kubenswrapper[4646]: I1203 10:58:57.896481 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" podUID="80cf56c0-7055-4e98-9fad-f728dd8b992a" containerName="route-controller-manager" containerID="cri-o://53310597ee480a18ec896d75cd1e1bb35d44fb4cd8f9ea335c8b84f38cd01943" gracePeriod=30 Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.167698 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.259390 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.349076 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-proxy-ca-bundles\") pod \"e51eada8-802d-4064-8705-2647ffa44189\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.349483 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wnzv\" (UniqueName: \"kubernetes.io/projected/e51eada8-802d-4064-8705-2647ffa44189-kube-api-access-8wnzv\") pod \"e51eada8-802d-4064-8705-2647ffa44189\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.349515 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-config\") pod \"e51eada8-802d-4064-8705-2647ffa44189\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.349541 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-client-ca\") pod \"e51eada8-802d-4064-8705-2647ffa44189\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.349572 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-578ds\" (UniqueName: \"kubernetes.io/projected/80cf56c0-7055-4e98-9fad-f728dd8b992a-kube-api-access-578ds\") pod \"80cf56c0-7055-4e98-9fad-f728dd8b992a\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.349610 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80cf56c0-7055-4e98-9fad-f728dd8b992a-client-ca\") pod \"80cf56c0-7055-4e98-9fad-f728dd8b992a\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.349670 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80cf56c0-7055-4e98-9fad-f728dd8b992a-serving-cert\") pod \"80cf56c0-7055-4e98-9fad-f728dd8b992a\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.349716 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80cf56c0-7055-4e98-9fad-f728dd8b992a-config\") pod \"80cf56c0-7055-4e98-9fad-f728dd8b992a\" (UID: \"80cf56c0-7055-4e98-9fad-f728dd8b992a\") " Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.349739 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e51eada8-802d-4064-8705-2647ffa44189-serving-cert\") pod \"e51eada8-802d-4064-8705-2647ffa44189\" (UID: \"e51eada8-802d-4064-8705-2647ffa44189\") " Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.349925 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e51eada8-802d-4064-8705-2647ffa44189" (UID: "e51eada8-802d-4064-8705-2647ffa44189"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.350473 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-config" (OuterVolumeSpecName: "config") pod "e51eada8-802d-4064-8705-2647ffa44189" (UID: "e51eada8-802d-4064-8705-2647ffa44189"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.350603 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-client-ca" (OuterVolumeSpecName: "client-ca") pod "e51eada8-802d-4064-8705-2647ffa44189" (UID: "e51eada8-802d-4064-8705-2647ffa44189"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.350795 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80cf56c0-7055-4e98-9fad-f728dd8b992a-client-ca" (OuterVolumeSpecName: "client-ca") pod "80cf56c0-7055-4e98-9fad-f728dd8b992a" (UID: "80cf56c0-7055-4e98-9fad-f728dd8b992a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.350835 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80cf56c0-7055-4e98-9fad-f728dd8b992a-config" (OuterVolumeSpecName: "config") pod "80cf56c0-7055-4e98-9fad-f728dd8b992a" (UID: "80cf56c0-7055-4e98-9fad-f728dd8b992a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.355008 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80cf56c0-7055-4e98-9fad-f728dd8b992a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "80cf56c0-7055-4e98-9fad-f728dd8b992a" (UID: "80cf56c0-7055-4e98-9fad-f728dd8b992a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.355034 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80cf56c0-7055-4e98-9fad-f728dd8b992a-kube-api-access-578ds" (OuterVolumeSpecName: "kube-api-access-578ds") pod "80cf56c0-7055-4e98-9fad-f728dd8b992a" (UID: "80cf56c0-7055-4e98-9fad-f728dd8b992a"). InnerVolumeSpecName "kube-api-access-578ds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.355018 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e51eada8-802d-4064-8705-2647ffa44189-kube-api-access-8wnzv" (OuterVolumeSpecName: "kube-api-access-8wnzv") pod "e51eada8-802d-4064-8705-2647ffa44189" (UID: "e51eada8-802d-4064-8705-2647ffa44189"). InnerVolumeSpecName "kube-api-access-8wnzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.355450 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e51eada8-802d-4064-8705-2647ffa44189-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e51eada8-802d-4064-8705-2647ffa44189" (UID: "e51eada8-802d-4064-8705-2647ffa44189"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.387270 4646 generic.go:334] "Generic (PLEG): container finished" podID="e51eada8-802d-4064-8705-2647ffa44189" containerID="f5b280f3f57a36058f5aa0b2f298d810688b521848dda8b6ce675fe80b058e23" exitCode=0 Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.387315 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" event={"ID":"e51eada8-802d-4064-8705-2647ffa44189","Type":"ContainerDied","Data":"f5b280f3f57a36058f5aa0b2f298d810688b521848dda8b6ce675fe80b058e23"} Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.387410 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" event={"ID":"e51eada8-802d-4064-8705-2647ffa44189","Type":"ContainerDied","Data":"dff929edc30561bd1de5123e47771db8e9af2fbabaea5811a77db03ccfbe01de"} Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.387460 4646 scope.go:117] "RemoveContainer" containerID="f5b280f3f57a36058f5aa0b2f298d810688b521848dda8b6ce675fe80b058e23" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.387689 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-czl94" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.388628 4646 generic.go:334] "Generic (PLEG): container finished" podID="80cf56c0-7055-4e98-9fad-f728dd8b992a" containerID="53310597ee480a18ec896d75cd1e1bb35d44fb4cd8f9ea335c8b84f38cd01943" exitCode=0 Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.388691 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" event={"ID":"80cf56c0-7055-4e98-9fad-f728dd8b992a","Type":"ContainerDied","Data":"53310597ee480a18ec896d75cd1e1bb35d44fb4cd8f9ea335c8b84f38cd01943"} Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.388709 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.388728 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj" event={"ID":"80cf56c0-7055-4e98-9fad-f728dd8b992a","Type":"ContainerDied","Data":"5282cf75c90fd44cc8ff0a10928f1f6686299264f2aca23aec56283b4f2ac1b1"} Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.407655 4646 scope.go:117] "RemoveContainer" containerID="f5b280f3f57a36058f5aa0b2f298d810688b521848dda8b6ce675fe80b058e23" Dec 03 10:58:58 crc kubenswrapper[4646]: E1203 10:58:58.408140 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5b280f3f57a36058f5aa0b2f298d810688b521848dda8b6ce675fe80b058e23\": container with ID starting with f5b280f3f57a36058f5aa0b2f298d810688b521848dda8b6ce675fe80b058e23 not found: ID does not exist" containerID="f5b280f3f57a36058f5aa0b2f298d810688b521848dda8b6ce675fe80b058e23" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.408180 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5b280f3f57a36058f5aa0b2f298d810688b521848dda8b6ce675fe80b058e23"} err="failed to get container status \"f5b280f3f57a36058f5aa0b2f298d810688b521848dda8b6ce675fe80b058e23\": rpc error: code = NotFound desc = could not find container \"f5b280f3f57a36058f5aa0b2f298d810688b521848dda8b6ce675fe80b058e23\": container with ID starting with f5b280f3f57a36058f5aa0b2f298d810688b521848dda8b6ce675fe80b058e23 not found: ID does not exist" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.408207 4646 scope.go:117] "RemoveContainer" containerID="53310597ee480a18ec896d75cd1e1bb35d44fb4cd8f9ea335c8b84f38cd01943" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.421224 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-czl94"] Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.423658 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-czl94"] Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.438085 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj"] Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.438134 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-plqjj"] Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.438240 4646 scope.go:117] "RemoveContainer" containerID="53310597ee480a18ec896d75cd1e1bb35d44fb4cd8f9ea335c8b84f38cd01943" Dec 03 10:58:58 crc kubenswrapper[4646]: E1203 10:58:58.439573 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53310597ee480a18ec896d75cd1e1bb35d44fb4cd8f9ea335c8b84f38cd01943\": container with ID starting with 53310597ee480a18ec896d75cd1e1bb35d44fb4cd8f9ea335c8b84f38cd01943 not found: ID does not exist" containerID="53310597ee480a18ec896d75cd1e1bb35d44fb4cd8f9ea335c8b84f38cd01943" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.439656 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53310597ee480a18ec896d75cd1e1bb35d44fb4cd8f9ea335c8b84f38cd01943"} err="failed to get container status \"53310597ee480a18ec896d75cd1e1bb35d44fb4cd8f9ea335c8b84f38cd01943\": rpc error: code = NotFound desc = could not find container \"53310597ee480a18ec896d75cd1e1bb35d44fb4cd8f9ea335c8b84f38cd01943\": container with ID starting with 53310597ee480a18ec896d75cd1e1bb35d44fb4cd8f9ea335c8b84f38cd01943 not found: ID does not exist" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.451466 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80cf56c0-7055-4e98-9fad-f728dd8b992a-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.451486 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e51eada8-802d-4064-8705-2647ffa44189-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.451496 4646 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.451505 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wnzv\" (UniqueName: \"kubernetes.io/projected/e51eada8-802d-4064-8705-2647ffa44189-kube-api-access-8wnzv\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.451514 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.451524 4646 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e51eada8-802d-4064-8705-2647ffa44189-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.451532 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-578ds\" (UniqueName: \"kubernetes.io/projected/80cf56c0-7055-4e98-9fad-f728dd8b992a-kube-api-access-578ds\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.451540 4646 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80cf56c0-7055-4e98-9fad-f728dd8b992a-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:58 crc kubenswrapper[4646]: I1203 10:58:58.451547 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80cf56c0-7055-4e98-9fad-f728dd8b992a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.093430 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.509742 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.624597 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-68ff5b985d-l4z5w"] Dec 03 10:58:59 crc kubenswrapper[4646]: E1203 10:58:59.624828 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="250f331c-acf3-4b28-b2a4-515a8cceb844" containerName="extract-content" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.624843 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="250f331c-acf3-4b28-b2a4-515a8cceb844" containerName="extract-content" Dec 03 10:58:59 crc kubenswrapper[4646]: E1203 10:58:59.624855 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="250f331c-acf3-4b28-b2a4-515a8cceb844" containerName="registry-server" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.624864 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="250f331c-acf3-4b28-b2a4-515a8cceb844" containerName="registry-server" Dec 03 10:58:59 crc kubenswrapper[4646]: E1203 10:58:59.624875 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13d8714-e047-4380-b095-7f5aac84f021" containerName="extract-utilities" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.624883 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13d8714-e047-4380-b095-7f5aac84f021" containerName="extract-utilities" Dec 03 10:58:59 crc kubenswrapper[4646]: E1203 10:58:59.624897 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1141f44-4444-47fb-878e-7a17db767b8d" containerName="extract-content" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.624905 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1141f44-4444-47fb-878e-7a17db767b8d" containerName="extract-content" Dec 03 10:58:59 crc kubenswrapper[4646]: E1203 10:58:59.624915 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20d9814-aa42-4109-9740-d0e5e6205a00" containerName="extract-content" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.624924 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20d9814-aa42-4109-9740-d0e5e6205a00" containerName="extract-content" Dec 03 10:58:59 crc kubenswrapper[4646]: E1203 10:58:59.624936 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1141f44-4444-47fb-878e-7a17db767b8d" containerName="registry-server" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.624944 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1141f44-4444-47fb-878e-7a17db767b8d" containerName="registry-server" Dec 03 10:58:59 crc kubenswrapper[4646]: E1203 10:58:59.624958 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="250f331c-acf3-4b28-b2a4-515a8cceb844" containerName="extract-utilities" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.624968 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="250f331c-acf3-4b28-b2a4-515a8cceb844" containerName="extract-utilities" Dec 03 10:58:59 crc kubenswrapper[4646]: E1203 10:58:59.624977 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20d9814-aa42-4109-9740-d0e5e6205a00" containerName="registry-server" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.624985 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20d9814-aa42-4109-9740-d0e5e6205a00" containerName="registry-server" Dec 03 10:58:59 crc kubenswrapper[4646]: E1203 10:58:59.624996 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80cf56c0-7055-4e98-9fad-f728dd8b992a" containerName="route-controller-manager" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.625005 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="80cf56c0-7055-4e98-9fad-f728dd8b992a" containerName="route-controller-manager" Dec 03 10:58:59 crc kubenswrapper[4646]: E1203 10:58:59.625018 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13d8714-e047-4380-b095-7f5aac84f021" containerName="extract-content" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.625026 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13d8714-e047-4380-b095-7f5aac84f021" containerName="extract-content" Dec 03 10:58:59 crc kubenswrapper[4646]: E1203 10:58:59.625038 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1141f44-4444-47fb-878e-7a17db767b8d" containerName="extract-utilities" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.625046 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1141f44-4444-47fb-878e-7a17db767b8d" containerName="extract-utilities" Dec 03 10:58:59 crc kubenswrapper[4646]: E1203 10:58:59.625056 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13d8714-e047-4380-b095-7f5aac84f021" containerName="registry-server" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.625064 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13d8714-e047-4380-b095-7f5aac84f021" containerName="registry-server" Dec 03 10:58:59 crc kubenswrapper[4646]: E1203 10:58:59.625074 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="143daef8-3687-4018-8f9f-2a675eac0909" containerName="marketplace-operator" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.625082 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="143daef8-3687-4018-8f9f-2a675eac0909" containerName="marketplace-operator" Dec 03 10:58:59 crc kubenswrapper[4646]: E1203 10:58:59.625092 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e51eada8-802d-4064-8705-2647ffa44189" containerName="controller-manager" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.625100 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="e51eada8-802d-4064-8705-2647ffa44189" containerName="controller-manager" Dec 03 10:58:59 crc kubenswrapper[4646]: E1203 10:58:59.625110 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20d9814-aa42-4109-9740-d0e5e6205a00" containerName="extract-utilities" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.625118 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20d9814-aa42-4109-9740-d0e5e6205a00" containerName="extract-utilities" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.625233 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="80cf56c0-7055-4e98-9fad-f728dd8b992a" containerName="route-controller-manager" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.625245 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="143daef8-3687-4018-8f9f-2a675eac0909" containerName="marketplace-operator" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.625256 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="e51eada8-802d-4064-8705-2647ffa44189" containerName="controller-manager" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.625268 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="b13d8714-e047-4380-b095-7f5aac84f021" containerName="registry-server" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.625280 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="250f331c-acf3-4b28-b2a4-515a8cceb844" containerName="registry-server" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.625291 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="d20d9814-aa42-4109-9740-d0e5e6205a00" containerName="registry-server" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.625303 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1141f44-4444-47fb-878e-7a17db767b8d" containerName="registry-server" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.625803 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.630124 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.630184 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.630398 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.631995 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.633005 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.638236 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.638866 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.665024 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-68ff5b985d-l4z5w"] Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.697301 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh"] Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.708371 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.713166 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh"] Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.724157 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.724299 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.724395 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.724435 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.724531 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.724562 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.773032 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-client-ca\") pod \"controller-manager-68ff5b985d-l4z5w\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.773094 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-serving-cert\") pod \"controller-manager-68ff5b985d-l4z5w\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.773157 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-proxy-ca-bundles\") pod \"controller-manager-68ff5b985d-l4z5w\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.773179 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-config\") pod \"controller-manager-68ff5b985d-l4z5w\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.773234 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgd2x\" (UniqueName: \"kubernetes.io/projected/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-kube-api-access-jgd2x\") pod \"controller-manager-68ff5b985d-l4z5w\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.853589 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80cf56c0-7055-4e98-9fad-f728dd8b992a" path="/var/lib/kubelet/pods/80cf56c0-7055-4e98-9fad-f728dd8b992a/volumes" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.854232 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e51eada8-802d-4064-8705-2647ffa44189" path="/var/lib/kubelet/pods/e51eada8-802d-4064-8705-2647ffa44189/volumes" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.874556 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgd2x\" (UniqueName: \"kubernetes.io/projected/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-kube-api-access-jgd2x\") pod \"controller-manager-68ff5b985d-l4z5w\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.874613 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvrrq\" (UniqueName: \"kubernetes.io/projected/ca7e2726-b8bb-45f5-af1d-e09456dfe033-kube-api-access-kvrrq\") pod \"route-controller-manager-8596db5b-qtckh\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.874653 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-client-ca\") pod \"controller-manager-68ff5b985d-l4z5w\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.874681 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-serving-cert\") pod \"controller-manager-68ff5b985d-l4z5w\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.874710 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ca7e2726-b8bb-45f5-af1d-e09456dfe033-client-ca\") pod \"route-controller-manager-8596db5b-qtckh\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.874810 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca7e2726-b8bb-45f5-af1d-e09456dfe033-config\") pod \"route-controller-manager-8596db5b-qtckh\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.874842 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-proxy-ca-bundles\") pod \"controller-manager-68ff5b985d-l4z5w\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.874871 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-config\") pod \"controller-manager-68ff5b985d-l4z5w\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.874906 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca7e2726-b8bb-45f5-af1d-e09456dfe033-serving-cert\") pod \"route-controller-manager-8596db5b-qtckh\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.875892 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-client-ca\") pod \"controller-manager-68ff5b985d-l4z5w\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.875946 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-proxy-ca-bundles\") pod \"controller-manager-68ff5b985d-l4z5w\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.876050 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-config\") pod \"controller-manager-68ff5b985d-l4z5w\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.881122 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-serving-cert\") pod \"controller-manager-68ff5b985d-l4z5w\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.897794 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgd2x\" (UniqueName: \"kubernetes.io/projected/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-kube-api-access-jgd2x\") pod \"controller-manager-68ff5b985d-l4z5w\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.942483 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.976438 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ca7e2726-b8bb-45f5-af1d-e09456dfe033-client-ca\") pod \"route-controller-manager-8596db5b-qtckh\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.976512 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca7e2726-b8bb-45f5-af1d-e09456dfe033-config\") pod \"route-controller-manager-8596db5b-qtckh\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.976558 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca7e2726-b8bb-45f5-af1d-e09456dfe033-serving-cert\") pod \"route-controller-manager-8596db5b-qtckh\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.977009 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvrrq\" (UniqueName: \"kubernetes.io/projected/ca7e2726-b8bb-45f5-af1d-e09456dfe033-kube-api-access-kvrrq\") pod \"route-controller-manager-8596db5b-qtckh\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.977414 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ca7e2726-b8bb-45f5-af1d-e09456dfe033-client-ca\") pod \"route-controller-manager-8596db5b-qtckh\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.978722 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca7e2726-b8bb-45f5-af1d-e09456dfe033-config\") pod \"route-controller-manager-8596db5b-qtckh\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.981217 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca7e2726-b8bb-45f5-af1d-e09456dfe033-serving-cert\") pod \"route-controller-manager-8596db5b-qtckh\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:58:59 crc kubenswrapper[4646]: I1203 10:58:59.996919 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvrrq\" (UniqueName: \"kubernetes.io/projected/ca7e2726-b8bb-45f5-af1d-e09456dfe033-kube-api-access-kvrrq\") pod \"route-controller-manager-8596db5b-qtckh\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:59:00 crc kubenswrapper[4646]: I1203 10:59:00.028858 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:59:00 crc kubenswrapper[4646]: I1203 10:59:00.143452 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-68ff5b985d-l4z5w"] Dec 03 10:59:00 crc kubenswrapper[4646]: I1203 10:59:00.274056 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh"] Dec 03 10:59:00 crc kubenswrapper[4646]: W1203 10:59:00.275081 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca7e2726_b8bb_45f5_af1d_e09456dfe033.slice/crio-a6690ddbf1910b4dd93d45a0b7967f160445fce2684f34feb63fc4581e3431c1 WatchSource:0}: Error finding container a6690ddbf1910b4dd93d45a0b7967f160445fce2684f34feb63fc4581e3431c1: Status 404 returned error can't find the container with id a6690ddbf1910b4dd93d45a0b7967f160445fce2684f34feb63fc4581e3431c1 Dec 03 10:59:00 crc kubenswrapper[4646]: I1203 10:59:00.401922 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" event={"ID":"ca7e2726-b8bb-45f5-af1d-e09456dfe033","Type":"ContainerStarted","Data":"a6690ddbf1910b4dd93d45a0b7967f160445fce2684f34feb63fc4581e3431c1"} Dec 03 10:59:00 crc kubenswrapper[4646]: I1203 10:59:00.403590 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" event={"ID":"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac","Type":"ContainerStarted","Data":"77473e8914d65c07b11983f568596b82ea84ee9e7fda60144c40af3dd164343e"} Dec 03 10:59:00 crc kubenswrapper[4646]: I1203 10:59:00.403619 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" event={"ID":"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac","Type":"ContainerStarted","Data":"75418963926a7b40b4b06c589a21a8eaae590c1f1bcd44f99ab4373d7a705e3a"} Dec 03 10:59:00 crc kubenswrapper[4646]: I1203 10:59:00.404587 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:59:00 crc kubenswrapper[4646]: I1203 10:59:00.409591 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:59:00 crc kubenswrapper[4646]: I1203 10:59:00.429855 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" podStartSLOduration=1.4298403149999999 podStartE2EDuration="1.429840315s" podCreationTimestamp="2025-12-03 10:58:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:59:00.426888182 +0000 UTC m=+316.889944327" watchObservedRunningTime="2025-12-03 10:59:00.429840315 +0000 UTC m=+316.892896460" Dec 03 10:59:00 crc kubenswrapper[4646]: I1203 10:59:00.966755 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 10:59:01 crc kubenswrapper[4646]: I1203 10:59:01.413039 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" event={"ID":"ca7e2726-b8bb-45f5-af1d-e09456dfe033","Type":"ContainerStarted","Data":"1b9fea0dc364ab19528a7a3c81f4db73d938707fff9d52c234752820f7e8af16"} Dec 03 10:59:01 crc kubenswrapper[4646]: I1203 10:59:01.414444 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:59:01 crc kubenswrapper[4646]: I1203 10:59:01.423480 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:59:01 crc kubenswrapper[4646]: I1203 10:59:01.468358 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" podStartSLOduration=2.468322755 podStartE2EDuration="2.468322755s" podCreationTimestamp="2025-12-03 10:58:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:59:01.445450345 +0000 UTC m=+317.908506500" watchObservedRunningTime="2025-12-03 10:59:01.468322755 +0000 UTC m=+317.931378900" Dec 03 10:59:02 crc kubenswrapper[4646]: I1203 10:59:02.617817 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.085198 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rz4r5"] Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.088652 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.092427 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rz4r5"] Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.096245 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.221501 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk4fq\" (UniqueName: \"kubernetes.io/projected/75816a6e-5fea-4212-964b-aafb2dce1385-kube-api-access-rk4fq\") pod \"certified-operators-rz4r5\" (UID: \"75816a6e-5fea-4212-964b-aafb2dce1385\") " pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.221552 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75816a6e-5fea-4212-964b-aafb2dce1385-utilities\") pod \"certified-operators-rz4r5\" (UID: \"75816a6e-5fea-4212-964b-aafb2dce1385\") " pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.221597 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75816a6e-5fea-4212-964b-aafb2dce1385-catalog-content\") pod \"certified-operators-rz4r5\" (UID: \"75816a6e-5fea-4212-964b-aafb2dce1385\") " pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.279390 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kbtw5"] Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.280500 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbtw5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.289285 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.291864 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kbtw5"] Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.322632 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk4fq\" (UniqueName: \"kubernetes.io/projected/75816a6e-5fea-4212-964b-aafb2dce1385-kube-api-access-rk4fq\") pod \"certified-operators-rz4r5\" (UID: \"75816a6e-5fea-4212-964b-aafb2dce1385\") " pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.322683 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75816a6e-5fea-4212-964b-aafb2dce1385-utilities\") pod \"certified-operators-rz4r5\" (UID: \"75816a6e-5fea-4212-964b-aafb2dce1385\") " pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.322746 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75816a6e-5fea-4212-964b-aafb2dce1385-catalog-content\") pod \"certified-operators-rz4r5\" (UID: \"75816a6e-5fea-4212-964b-aafb2dce1385\") " pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.323370 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75816a6e-5fea-4212-964b-aafb2dce1385-catalog-content\") pod \"certified-operators-rz4r5\" (UID: \"75816a6e-5fea-4212-964b-aafb2dce1385\") " pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.323394 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75816a6e-5fea-4212-964b-aafb2dce1385-utilities\") pod \"certified-operators-rz4r5\" (UID: \"75816a6e-5fea-4212-964b-aafb2dce1385\") " pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.340123 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk4fq\" (UniqueName: \"kubernetes.io/projected/75816a6e-5fea-4212-964b-aafb2dce1385-kube-api-access-rk4fq\") pod \"certified-operators-rz4r5\" (UID: \"75816a6e-5fea-4212-964b-aafb2dce1385\") " pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.404925 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.423880 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ccf8b15-522e-4425-9ce3-8ba8721454b7-utilities\") pod \"community-operators-kbtw5\" (UID: \"9ccf8b15-522e-4425-9ce3-8ba8721454b7\") " pod="openshift-marketplace/community-operators-kbtw5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.423970 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ccf8b15-522e-4425-9ce3-8ba8721454b7-catalog-content\") pod \"community-operators-kbtw5\" (UID: \"9ccf8b15-522e-4425-9ce3-8ba8721454b7\") " pod="openshift-marketplace/community-operators-kbtw5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.424030 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtd7w\" (UniqueName: \"kubernetes.io/projected/9ccf8b15-522e-4425-9ce3-8ba8721454b7-kube-api-access-wtd7w\") pod \"community-operators-kbtw5\" (UID: \"9ccf8b15-522e-4425-9ce3-8ba8721454b7\") " pod="openshift-marketplace/community-operators-kbtw5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.525071 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ccf8b15-522e-4425-9ce3-8ba8721454b7-catalog-content\") pod \"community-operators-kbtw5\" (UID: \"9ccf8b15-522e-4425-9ce3-8ba8721454b7\") " pod="openshift-marketplace/community-operators-kbtw5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.525148 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtd7w\" (UniqueName: \"kubernetes.io/projected/9ccf8b15-522e-4425-9ce3-8ba8721454b7-kube-api-access-wtd7w\") pod \"community-operators-kbtw5\" (UID: \"9ccf8b15-522e-4425-9ce3-8ba8721454b7\") " pod="openshift-marketplace/community-operators-kbtw5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.525178 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ccf8b15-522e-4425-9ce3-8ba8721454b7-utilities\") pod \"community-operators-kbtw5\" (UID: \"9ccf8b15-522e-4425-9ce3-8ba8721454b7\") " pod="openshift-marketplace/community-operators-kbtw5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.525645 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ccf8b15-522e-4425-9ce3-8ba8721454b7-utilities\") pod \"community-operators-kbtw5\" (UID: \"9ccf8b15-522e-4425-9ce3-8ba8721454b7\") " pod="openshift-marketplace/community-operators-kbtw5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.530145 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ccf8b15-522e-4425-9ce3-8ba8721454b7-catalog-content\") pod \"community-operators-kbtw5\" (UID: \"9ccf8b15-522e-4425-9ce3-8ba8721454b7\") " pod="openshift-marketplace/community-operators-kbtw5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.549370 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtd7w\" (UniqueName: \"kubernetes.io/projected/9ccf8b15-522e-4425-9ce3-8ba8721454b7-kube-api-access-wtd7w\") pod \"community-operators-kbtw5\" (UID: \"9ccf8b15-522e-4425-9ce3-8ba8721454b7\") " pod="openshift-marketplace/community-operators-kbtw5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.594943 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbtw5" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.814604 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rz4r5"] Dec 03 10:59:33 crc kubenswrapper[4646]: W1203 10:59:33.821645 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75816a6e_5fea_4212_964b_aafb2dce1385.slice/crio-dc5b0bd19899617f840dc714cbefc1ed9700627375a2e3360f65313030f65cdc WatchSource:0}: Error finding container dc5b0bd19899617f840dc714cbefc1ed9700627375a2e3360f65313030f65cdc: Status 404 returned error can't find the container with id dc5b0bd19899617f840dc714cbefc1ed9700627375a2e3360f65313030f65cdc Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.986994 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-sd54l"] Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.987564 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:33 crc kubenswrapper[4646]: I1203 10:59:33.999188 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-sd54l"] Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.034604 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kbtw5"] Dec 03 10:59:34 crc kubenswrapper[4646]: W1203 10:59:34.042705 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ccf8b15_522e_4425_9ce3_8ba8721454b7.slice/crio-4a0cb3257d0ee89f389868314fd6e5747c1532537bfe7170a37a88bfe2298a7d WatchSource:0}: Error finding container 4a0cb3257d0ee89f389868314fd6e5747c1532537bfe7170a37a88bfe2298a7d: Status 404 returned error can't find the container with id 4a0cb3257d0ee89f389868314fd6e5747c1532537bfe7170a37a88bfe2298a7d Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.131246 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6f33ec80-8950-4ddc-bad5-e23cd63a2881-ca-trust-extracted\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.131310 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6f33ec80-8950-4ddc-bad5-e23cd63a2881-bound-sa-token\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.131352 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6f33ec80-8950-4ddc-bad5-e23cd63a2881-registry-certificates\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.131392 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6f33ec80-8950-4ddc-bad5-e23cd63a2881-installation-pull-secrets\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.131413 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6f33ec80-8950-4ddc-bad5-e23cd63a2881-registry-tls\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.131493 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwsbb\" (UniqueName: \"kubernetes.io/projected/6f33ec80-8950-4ddc-bad5-e23cd63a2881-kube-api-access-dwsbb\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.131571 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.131630 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f33ec80-8950-4ddc-bad5-e23cd63a2881-trusted-ca\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.153982 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.233033 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6f33ec80-8950-4ddc-bad5-e23cd63a2881-bound-sa-token\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.233073 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6f33ec80-8950-4ddc-bad5-e23cd63a2881-registry-certificates\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.233113 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6f33ec80-8950-4ddc-bad5-e23cd63a2881-installation-pull-secrets\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.233137 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6f33ec80-8950-4ddc-bad5-e23cd63a2881-registry-tls\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.233158 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwsbb\" (UniqueName: \"kubernetes.io/projected/6f33ec80-8950-4ddc-bad5-e23cd63a2881-kube-api-access-dwsbb\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.233193 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f33ec80-8950-4ddc-bad5-e23cd63a2881-trusted-ca\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.233209 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6f33ec80-8950-4ddc-bad5-e23cd63a2881-ca-trust-extracted\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.233589 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6f33ec80-8950-4ddc-bad5-e23cd63a2881-ca-trust-extracted\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.234708 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6f33ec80-8950-4ddc-bad5-e23cd63a2881-registry-certificates\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.236260 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f33ec80-8950-4ddc-bad5-e23cd63a2881-trusted-ca\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.239112 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6f33ec80-8950-4ddc-bad5-e23cd63a2881-installation-pull-secrets\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.239498 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6f33ec80-8950-4ddc-bad5-e23cd63a2881-registry-tls\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.255939 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwsbb\" (UniqueName: \"kubernetes.io/projected/6f33ec80-8950-4ddc-bad5-e23cd63a2881-kube-api-access-dwsbb\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.257910 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6f33ec80-8950-4ddc-bad5-e23cd63a2881-bound-sa-token\") pod \"image-registry-66df7c8f76-sd54l\" (UID: \"6f33ec80-8950-4ddc-bad5-e23cd63a2881\") " pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.303421 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.614890 4646 generic.go:334] "Generic (PLEG): container finished" podID="9ccf8b15-522e-4425-9ce3-8ba8721454b7" containerID="13769f559cfce58c5111d2e8b3d98d8e4f9a1d50d4143fa9b37a12afdfa6b233" exitCode=0 Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.614969 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbtw5" event={"ID":"9ccf8b15-522e-4425-9ce3-8ba8721454b7","Type":"ContainerDied","Data":"13769f559cfce58c5111d2e8b3d98d8e4f9a1d50d4143fa9b37a12afdfa6b233"} Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.615190 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbtw5" event={"ID":"9ccf8b15-522e-4425-9ce3-8ba8721454b7","Type":"ContainerStarted","Data":"4a0cb3257d0ee89f389868314fd6e5747c1532537bfe7170a37a88bfe2298a7d"} Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.637020 4646 generic.go:334] "Generic (PLEG): container finished" podID="75816a6e-5fea-4212-964b-aafb2dce1385" containerID="b0c2880f23237eb03391a50ab85c8c308f9d883e054bfe555390c47e792535fe" exitCode=0 Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.637101 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rz4r5" event={"ID":"75816a6e-5fea-4212-964b-aafb2dce1385","Type":"ContainerDied","Data":"b0c2880f23237eb03391a50ab85c8c308f9d883e054bfe555390c47e792535fe"} Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.637150 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rz4r5" event={"ID":"75816a6e-5fea-4212-964b-aafb2dce1385","Type":"ContainerStarted","Data":"dc5b0bd19899617f840dc714cbefc1ed9700627375a2e3360f65313030f65cdc"} Dec 03 10:59:34 crc kubenswrapper[4646]: I1203 10:59:34.762853 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-sd54l"] Dec 03 10:59:34 crc kubenswrapper[4646]: W1203 10:59:34.772195 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f33ec80_8950_4ddc_bad5_e23cd63a2881.slice/crio-e049c96112cb1830b8b908d11592da2656dc55609918cdf9704ac9a79378ddfa WatchSource:0}: Error finding container e049c96112cb1830b8b908d11592da2656dc55609918cdf9704ac9a79378ddfa: Status 404 returned error can't find the container with id e049c96112cb1830b8b908d11592da2656dc55609918cdf9704ac9a79378ddfa Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.496734 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dtn6s"] Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.500675 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtn6s" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.504988 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.506509 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtn6s"] Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.653142 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5394e3ff-06b8-4220-81c1-21475d7769fa-catalog-content\") pod \"redhat-marketplace-dtn6s\" (UID: \"5394e3ff-06b8-4220-81c1-21475d7769fa\") " pod="openshift-marketplace/redhat-marketplace-dtn6s" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.653184 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5394e3ff-06b8-4220-81c1-21475d7769fa-utilities\") pod \"redhat-marketplace-dtn6s\" (UID: \"5394e3ff-06b8-4220-81c1-21475d7769fa\") " pod="openshift-marketplace/redhat-marketplace-dtn6s" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.653250 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56cr2\" (UniqueName: \"kubernetes.io/projected/5394e3ff-06b8-4220-81c1-21475d7769fa-kube-api-access-56cr2\") pod \"redhat-marketplace-dtn6s\" (UID: \"5394e3ff-06b8-4220-81c1-21475d7769fa\") " pod="openshift-marketplace/redhat-marketplace-dtn6s" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.653580 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbtw5" event={"ID":"9ccf8b15-522e-4425-9ce3-8ba8721454b7","Type":"ContainerStarted","Data":"bc05871670b6ad184a8a8d6f760e296d4b1ca4ac44fa4c4fddd22977a71923a8"} Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.654821 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" event={"ID":"6f33ec80-8950-4ddc-bad5-e23cd63a2881","Type":"ContainerStarted","Data":"6248cd452e9831b59d33ae53322f6a58e694fec41afdcb8f48508851452da1c8"} Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.654860 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" event={"ID":"6f33ec80-8950-4ddc-bad5-e23cd63a2881","Type":"ContainerStarted","Data":"e049c96112cb1830b8b908d11592da2656dc55609918cdf9704ac9a79378ddfa"} Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.654948 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.656183 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rz4r5" event={"ID":"75816a6e-5fea-4212-964b-aafb2dce1385","Type":"ContainerStarted","Data":"489acfe2db031321c437c8680630e180c844e3616468958da33b16fa88ae8053"} Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.683739 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-82b96"] Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.684641 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-82b96" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.687578 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.700591 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-82b96"] Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.718997 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" podStartSLOduration=2.7189781 podStartE2EDuration="2.7189781s" podCreationTimestamp="2025-12-03 10:59:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:59:35.71834758 +0000 UTC m=+352.181403715" watchObservedRunningTime="2025-12-03 10:59:35.7189781 +0000 UTC m=+352.182034225" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.754920 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56cr2\" (UniqueName: \"kubernetes.io/projected/5394e3ff-06b8-4220-81c1-21475d7769fa-kube-api-access-56cr2\") pod \"redhat-marketplace-dtn6s\" (UID: \"5394e3ff-06b8-4220-81c1-21475d7769fa\") " pod="openshift-marketplace/redhat-marketplace-dtn6s" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.755247 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5394e3ff-06b8-4220-81c1-21475d7769fa-catalog-content\") pod \"redhat-marketplace-dtn6s\" (UID: \"5394e3ff-06b8-4220-81c1-21475d7769fa\") " pod="openshift-marketplace/redhat-marketplace-dtn6s" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.755742 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5394e3ff-06b8-4220-81c1-21475d7769fa-utilities\") pod \"redhat-marketplace-dtn6s\" (UID: \"5394e3ff-06b8-4220-81c1-21475d7769fa\") " pod="openshift-marketplace/redhat-marketplace-dtn6s" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.755923 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5394e3ff-06b8-4220-81c1-21475d7769fa-catalog-content\") pod \"redhat-marketplace-dtn6s\" (UID: \"5394e3ff-06b8-4220-81c1-21475d7769fa\") " pod="openshift-marketplace/redhat-marketplace-dtn6s" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.756050 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5394e3ff-06b8-4220-81c1-21475d7769fa-utilities\") pod \"redhat-marketplace-dtn6s\" (UID: \"5394e3ff-06b8-4220-81c1-21475d7769fa\") " pod="openshift-marketplace/redhat-marketplace-dtn6s" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.773978 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56cr2\" (UniqueName: \"kubernetes.io/projected/5394e3ff-06b8-4220-81c1-21475d7769fa-kube-api-access-56cr2\") pod \"redhat-marketplace-dtn6s\" (UID: \"5394e3ff-06b8-4220-81c1-21475d7769fa\") " pod="openshift-marketplace/redhat-marketplace-dtn6s" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.823803 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtn6s" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.856764 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x2kf\" (UniqueName: \"kubernetes.io/projected/0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e-kube-api-access-4x2kf\") pod \"redhat-operators-82b96\" (UID: \"0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e\") " pod="openshift-marketplace/redhat-operators-82b96" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.856867 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e-catalog-content\") pod \"redhat-operators-82b96\" (UID: \"0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e\") " pod="openshift-marketplace/redhat-operators-82b96" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.856905 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e-utilities\") pod \"redhat-operators-82b96\" (UID: \"0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e\") " pod="openshift-marketplace/redhat-operators-82b96" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.957968 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e-catalog-content\") pod \"redhat-operators-82b96\" (UID: \"0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e\") " pod="openshift-marketplace/redhat-operators-82b96" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.958697 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e-utilities\") pod \"redhat-operators-82b96\" (UID: \"0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e\") " pod="openshift-marketplace/redhat-operators-82b96" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.958781 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x2kf\" (UniqueName: \"kubernetes.io/projected/0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e-kube-api-access-4x2kf\") pod \"redhat-operators-82b96\" (UID: \"0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e\") " pod="openshift-marketplace/redhat-operators-82b96" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.960566 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e-catalog-content\") pod \"redhat-operators-82b96\" (UID: \"0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e\") " pod="openshift-marketplace/redhat-operators-82b96" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.960700 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e-utilities\") pod \"redhat-operators-82b96\" (UID: \"0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e\") " pod="openshift-marketplace/redhat-operators-82b96" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.983530 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x2kf\" (UniqueName: \"kubernetes.io/projected/0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e-kube-api-access-4x2kf\") pod \"redhat-operators-82b96\" (UID: \"0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e\") " pod="openshift-marketplace/redhat-operators-82b96" Dec 03 10:59:35 crc kubenswrapper[4646]: I1203 10:59:35.996102 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-82b96" Dec 03 10:59:36 crc kubenswrapper[4646]: I1203 10:59:36.028924 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtn6s"] Dec 03 10:59:36 crc kubenswrapper[4646]: I1203 10:59:36.430027 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-82b96"] Dec 03 10:59:36 crc kubenswrapper[4646]: W1203 10:59:36.438069 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0bc4f0c6_82a4_4b75_bf15_3f4e5a4f360e.slice/crio-812477b950a76f0100d7803b4f9ad677702d4d32b35feccd6e6c7fe351208131 WatchSource:0}: Error finding container 812477b950a76f0100d7803b4f9ad677702d4d32b35feccd6e6c7fe351208131: Status 404 returned error can't find the container with id 812477b950a76f0100d7803b4f9ad677702d4d32b35feccd6e6c7fe351208131 Dec 03 10:59:36 crc kubenswrapper[4646]: I1203 10:59:36.666191 4646 generic.go:334] "Generic (PLEG): container finished" podID="5394e3ff-06b8-4220-81c1-21475d7769fa" containerID="f76d578448bdd106689dafc0ccc6ae2c0836a283a38b15d83e7e49db0ee5651e" exitCode=0 Dec 03 10:59:36 crc kubenswrapper[4646]: I1203 10:59:36.666434 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtn6s" event={"ID":"5394e3ff-06b8-4220-81c1-21475d7769fa","Type":"ContainerDied","Data":"f76d578448bdd106689dafc0ccc6ae2c0836a283a38b15d83e7e49db0ee5651e"} Dec 03 10:59:36 crc kubenswrapper[4646]: I1203 10:59:36.666496 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtn6s" event={"ID":"5394e3ff-06b8-4220-81c1-21475d7769fa","Type":"ContainerStarted","Data":"4dc4d1d052f6f46681e7759429be10b64380f228254868bf8f6100d61707984d"} Dec 03 10:59:36 crc kubenswrapper[4646]: I1203 10:59:36.668887 4646 generic.go:334] "Generic (PLEG): container finished" podID="0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e" containerID="a1d2af5101a34baf30ace68d20e2b3e327d7e93345cad748f870420727ee2159" exitCode=0 Dec 03 10:59:36 crc kubenswrapper[4646]: I1203 10:59:36.669103 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82b96" event={"ID":"0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e","Type":"ContainerDied","Data":"a1d2af5101a34baf30ace68d20e2b3e327d7e93345cad748f870420727ee2159"} Dec 03 10:59:36 crc kubenswrapper[4646]: I1203 10:59:36.669139 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82b96" event={"ID":"0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e","Type":"ContainerStarted","Data":"812477b950a76f0100d7803b4f9ad677702d4d32b35feccd6e6c7fe351208131"} Dec 03 10:59:36 crc kubenswrapper[4646]: I1203 10:59:36.673692 4646 generic.go:334] "Generic (PLEG): container finished" podID="75816a6e-5fea-4212-964b-aafb2dce1385" containerID="489acfe2db031321c437c8680630e180c844e3616468958da33b16fa88ae8053" exitCode=0 Dec 03 10:59:36 crc kubenswrapper[4646]: I1203 10:59:36.673787 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rz4r5" event={"ID":"75816a6e-5fea-4212-964b-aafb2dce1385","Type":"ContainerDied","Data":"489acfe2db031321c437c8680630e180c844e3616468958da33b16fa88ae8053"} Dec 03 10:59:36 crc kubenswrapper[4646]: I1203 10:59:36.687807 4646 generic.go:334] "Generic (PLEG): container finished" podID="9ccf8b15-522e-4425-9ce3-8ba8721454b7" containerID="bc05871670b6ad184a8a8d6f760e296d4b1ca4ac44fa4c4fddd22977a71923a8" exitCode=0 Dec 03 10:59:36 crc kubenswrapper[4646]: I1203 10:59:36.691953 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbtw5" event={"ID":"9ccf8b15-522e-4425-9ce3-8ba8721454b7","Type":"ContainerDied","Data":"bc05871670b6ad184a8a8d6f760e296d4b1ca4ac44fa4c4fddd22977a71923a8"} Dec 03 10:59:37 crc kubenswrapper[4646]: E1203 10:59:37.544951 4646 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5394e3ff_06b8_4220_81c1_21475d7769fa.slice/crio-conmon-321e3a9e3bf6646537c520ae75deaf4d05e6298bc9ae7d0f9f8075670bd44ef7.scope\": RecentStats: unable to find data in memory cache]" Dec 03 10:59:37 crc kubenswrapper[4646]: I1203 10:59:37.702330 4646 generic.go:334] "Generic (PLEG): container finished" podID="5394e3ff-06b8-4220-81c1-21475d7769fa" containerID="321e3a9e3bf6646537c520ae75deaf4d05e6298bc9ae7d0f9f8075670bd44ef7" exitCode=0 Dec 03 10:59:37 crc kubenswrapper[4646]: I1203 10:59:37.702389 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtn6s" event={"ID":"5394e3ff-06b8-4220-81c1-21475d7769fa","Type":"ContainerDied","Data":"321e3a9e3bf6646537c520ae75deaf4d05e6298bc9ae7d0f9f8075670bd44ef7"} Dec 03 10:59:37 crc kubenswrapper[4646]: I1203 10:59:37.704768 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82b96" event={"ID":"0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e","Type":"ContainerStarted","Data":"9173f8427131bc0f142979f5bc7b4f92db2190c2aadd87809fbb06714578ef3d"} Dec 03 10:59:37 crc kubenswrapper[4646]: I1203 10:59:37.709930 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rz4r5" event={"ID":"75816a6e-5fea-4212-964b-aafb2dce1385","Type":"ContainerStarted","Data":"9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b"} Dec 03 10:59:37 crc kubenswrapper[4646]: I1203 10:59:37.711872 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbtw5" event={"ID":"9ccf8b15-522e-4425-9ce3-8ba8721454b7","Type":"ContainerStarted","Data":"e87b3bd3cbb0f757866f025896cfdf12a167f50949c61a60583e28e04225a752"} Dec 03 10:59:37 crc kubenswrapper[4646]: I1203 10:59:37.746493 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kbtw5" podStartSLOduration=2.176025819 podStartE2EDuration="4.746476252s" podCreationTimestamp="2025-12-03 10:59:33 +0000 UTC" firstStartedPulling="2025-12-03 10:59:34.62168923 +0000 UTC m=+351.084745385" lastFinishedPulling="2025-12-03 10:59:37.192139633 +0000 UTC m=+353.655195818" observedRunningTime="2025-12-03 10:59:37.743844209 +0000 UTC m=+354.206900344" watchObservedRunningTime="2025-12-03 10:59:37.746476252 +0000 UTC m=+354.209532387" Dec 03 10:59:37 crc kubenswrapper[4646]: I1203 10:59:37.785515 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rz4r5" podStartSLOduration=2.320906361 podStartE2EDuration="4.785494731s" podCreationTimestamp="2025-12-03 10:59:33 +0000 UTC" firstStartedPulling="2025-12-03 10:59:34.642302479 +0000 UTC m=+351.105358654" lastFinishedPulling="2025-12-03 10:59:37.106890859 +0000 UTC m=+353.569947024" observedRunningTime="2025-12-03 10:59:37.768598399 +0000 UTC m=+354.231654534" watchObservedRunningTime="2025-12-03 10:59:37.785494731 +0000 UTC m=+354.248550866" Dec 03 10:59:37 crc kubenswrapper[4646]: I1203 10:59:37.826014 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-68ff5b985d-l4z5w"] Dec 03 10:59:37 crc kubenswrapper[4646]: I1203 10:59:37.826238 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" podUID="62d1d878-ed0e-4901-bd0b-5aa2d8f05bac" containerName="controller-manager" containerID="cri-o://77473e8914d65c07b11983f568596b82ea84ee9e7fda60144c40af3dd164343e" gracePeriod=30 Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.269581 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.397986 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-client-ca\") pod \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.398036 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-config\") pod \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.398068 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgd2x\" (UniqueName: \"kubernetes.io/projected/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-kube-api-access-jgd2x\") pod \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.398105 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-proxy-ca-bundles\") pod \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.398160 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-serving-cert\") pod \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\" (UID: \"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac\") " Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.399833 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-config" (OuterVolumeSpecName: "config") pod "62d1d878-ed0e-4901-bd0b-5aa2d8f05bac" (UID: "62d1d878-ed0e-4901-bd0b-5aa2d8f05bac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.400067 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-client-ca" (OuterVolumeSpecName: "client-ca") pod "62d1d878-ed0e-4901-bd0b-5aa2d8f05bac" (UID: "62d1d878-ed0e-4901-bd0b-5aa2d8f05bac"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.400302 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "62d1d878-ed0e-4901-bd0b-5aa2d8f05bac" (UID: "62d1d878-ed0e-4901-bd0b-5aa2d8f05bac"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.407611 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-kube-api-access-jgd2x" (OuterVolumeSpecName: "kube-api-access-jgd2x") pod "62d1d878-ed0e-4901-bd0b-5aa2d8f05bac" (UID: "62d1d878-ed0e-4901-bd0b-5aa2d8f05bac"). InnerVolumeSpecName "kube-api-access-jgd2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.407743 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "62d1d878-ed0e-4901-bd0b-5aa2d8f05bac" (UID: "62d1d878-ed0e-4901-bd0b-5aa2d8f05bac"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.499854 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgd2x\" (UniqueName: \"kubernetes.io/projected/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-kube-api-access-jgd2x\") on node \"crc\" DevicePath \"\"" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.499892 4646 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.499901 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.499910 4646 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.499918 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.720367 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtn6s" event={"ID":"5394e3ff-06b8-4220-81c1-21475d7769fa","Type":"ContainerStarted","Data":"fdefe09decb860d07a1bbc8a6880fd8bafa746165cfde20ecdd7a9b62a6ad450"} Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.721780 4646 generic.go:334] "Generic (PLEG): container finished" podID="62d1d878-ed0e-4901-bd0b-5aa2d8f05bac" containerID="77473e8914d65c07b11983f568596b82ea84ee9e7fda60144c40af3dd164343e" exitCode=0 Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.721808 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" event={"ID":"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac","Type":"ContainerDied","Data":"77473e8914d65c07b11983f568596b82ea84ee9e7fda60144c40af3dd164343e"} Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.721830 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" event={"ID":"62d1d878-ed0e-4901-bd0b-5aa2d8f05bac","Type":"ContainerDied","Data":"75418963926a7b40b4b06c589a21a8eaae590c1f1bcd44f99ab4373d7a705e3a"} Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.721847 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-68ff5b985d-l4z5w" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.721847 4646 scope.go:117] "RemoveContainer" containerID="77473e8914d65c07b11983f568596b82ea84ee9e7fda60144c40af3dd164343e" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.723525 4646 generic.go:334] "Generic (PLEG): container finished" podID="0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e" containerID="9173f8427131bc0f142979f5bc7b4f92db2190c2aadd87809fbb06714578ef3d" exitCode=0 Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.723671 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82b96" event={"ID":"0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e","Type":"ContainerDied","Data":"9173f8427131bc0f142979f5bc7b4f92db2190c2aadd87809fbb06714578ef3d"} Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.742025 4646 scope.go:117] "RemoveContainer" containerID="77473e8914d65c07b11983f568596b82ea84ee9e7fda60144c40af3dd164343e" Dec 03 10:59:38 crc kubenswrapper[4646]: E1203 10:59:38.742395 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77473e8914d65c07b11983f568596b82ea84ee9e7fda60144c40af3dd164343e\": container with ID starting with 77473e8914d65c07b11983f568596b82ea84ee9e7fda60144c40af3dd164343e not found: ID does not exist" containerID="77473e8914d65c07b11983f568596b82ea84ee9e7fda60144c40af3dd164343e" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.742426 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77473e8914d65c07b11983f568596b82ea84ee9e7fda60144c40af3dd164343e"} err="failed to get container status \"77473e8914d65c07b11983f568596b82ea84ee9e7fda60144c40af3dd164343e\": rpc error: code = NotFound desc = could not find container \"77473e8914d65c07b11983f568596b82ea84ee9e7fda60144c40af3dd164343e\": container with ID starting with 77473e8914d65c07b11983f568596b82ea84ee9e7fda60144c40af3dd164343e not found: ID does not exist" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.749662 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dtn6s" podStartSLOduration=2.286249704 podStartE2EDuration="3.749647809s" podCreationTimestamp="2025-12-03 10:59:35 +0000 UTC" firstStartedPulling="2025-12-03 10:59:36.668632174 +0000 UTC m=+353.131688319" lastFinishedPulling="2025-12-03 10:59:38.132030289 +0000 UTC m=+354.595086424" observedRunningTime="2025-12-03 10:59:38.746073787 +0000 UTC m=+355.209129932" watchObservedRunningTime="2025-12-03 10:59:38.749647809 +0000 UTC m=+355.212703944" Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.760871 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-68ff5b985d-l4z5w"] Dec 03 10:59:38 crc kubenswrapper[4646]: I1203 10:59:38.763515 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-68ff5b985d-l4z5w"] Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.728811 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82b96" event={"ID":"0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e","Type":"ContainerStarted","Data":"734f9471aa4e1e4fa3bd8390ce2e6b255b70a4e79cc62ab456a44867930d6bc3"} Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.751997 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-82b96" podStartSLOduration=2.211660238 podStartE2EDuration="4.751972941s" podCreationTimestamp="2025-12-03 10:59:35 +0000 UTC" firstStartedPulling="2025-12-03 10:59:36.672061712 +0000 UTC m=+353.135117857" lastFinishedPulling="2025-12-03 10:59:39.212374425 +0000 UTC m=+355.675430560" observedRunningTime="2025-12-03 10:59:39.749261156 +0000 UTC m=+356.212317301" watchObservedRunningTime="2025-12-03 10:59:39.751972941 +0000 UTC m=+356.215029096" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.772030 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-79f967984f-csln4"] Dec 03 10:59:39 crc kubenswrapper[4646]: E1203 10:59:39.772231 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d1d878-ed0e-4901-bd0b-5aa2d8f05bac" containerName="controller-manager" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.772242 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d1d878-ed0e-4901-bd0b-5aa2d8f05bac" containerName="controller-manager" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.772329 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d1d878-ed0e-4901-bd0b-5aa2d8f05bac" containerName="controller-manager" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.772685 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.774516 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.774679 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.776578 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.776707 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.776731 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.777682 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.783237 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.788945 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-79f967984f-csln4"] Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.855427 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62d1d878-ed0e-4901-bd0b-5aa2d8f05bac" path="/var/lib/kubelet/pods/62d1d878-ed0e-4901-bd0b-5aa2d8f05bac/volumes" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.915235 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c77a1b29-f590-4342-a62a-1b3fe00863b7-proxy-ca-bundles\") pod \"controller-manager-79f967984f-csln4\" (UID: \"c77a1b29-f590-4342-a62a-1b3fe00863b7\") " pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.915306 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c77a1b29-f590-4342-a62a-1b3fe00863b7-client-ca\") pod \"controller-manager-79f967984f-csln4\" (UID: \"c77a1b29-f590-4342-a62a-1b3fe00863b7\") " pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.915575 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c77a1b29-f590-4342-a62a-1b3fe00863b7-serving-cert\") pod \"controller-manager-79f967984f-csln4\" (UID: \"c77a1b29-f590-4342-a62a-1b3fe00863b7\") " pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.915700 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph5s9\" (UniqueName: \"kubernetes.io/projected/c77a1b29-f590-4342-a62a-1b3fe00863b7-kube-api-access-ph5s9\") pod \"controller-manager-79f967984f-csln4\" (UID: \"c77a1b29-f590-4342-a62a-1b3fe00863b7\") " pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:39 crc kubenswrapper[4646]: I1203 10:59:39.915846 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c77a1b29-f590-4342-a62a-1b3fe00863b7-config\") pod \"controller-manager-79f967984f-csln4\" (UID: \"c77a1b29-f590-4342-a62a-1b3fe00863b7\") " pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:40 crc kubenswrapper[4646]: I1203 10:59:40.016719 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c77a1b29-f590-4342-a62a-1b3fe00863b7-serving-cert\") pod \"controller-manager-79f967984f-csln4\" (UID: \"c77a1b29-f590-4342-a62a-1b3fe00863b7\") " pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:40 crc kubenswrapper[4646]: I1203 10:59:40.016752 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph5s9\" (UniqueName: \"kubernetes.io/projected/c77a1b29-f590-4342-a62a-1b3fe00863b7-kube-api-access-ph5s9\") pod \"controller-manager-79f967984f-csln4\" (UID: \"c77a1b29-f590-4342-a62a-1b3fe00863b7\") " pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:40 crc kubenswrapper[4646]: I1203 10:59:40.016828 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c77a1b29-f590-4342-a62a-1b3fe00863b7-config\") pod \"controller-manager-79f967984f-csln4\" (UID: \"c77a1b29-f590-4342-a62a-1b3fe00863b7\") " pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:40 crc kubenswrapper[4646]: I1203 10:59:40.016866 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c77a1b29-f590-4342-a62a-1b3fe00863b7-proxy-ca-bundles\") pod \"controller-manager-79f967984f-csln4\" (UID: \"c77a1b29-f590-4342-a62a-1b3fe00863b7\") " pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:40 crc kubenswrapper[4646]: I1203 10:59:40.016915 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c77a1b29-f590-4342-a62a-1b3fe00863b7-client-ca\") pod \"controller-manager-79f967984f-csln4\" (UID: \"c77a1b29-f590-4342-a62a-1b3fe00863b7\") " pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:40 crc kubenswrapper[4646]: I1203 10:59:40.018003 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c77a1b29-f590-4342-a62a-1b3fe00863b7-proxy-ca-bundles\") pod \"controller-manager-79f967984f-csln4\" (UID: \"c77a1b29-f590-4342-a62a-1b3fe00863b7\") " pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:40 crc kubenswrapper[4646]: I1203 10:59:40.018465 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c77a1b29-f590-4342-a62a-1b3fe00863b7-client-ca\") pod \"controller-manager-79f967984f-csln4\" (UID: \"c77a1b29-f590-4342-a62a-1b3fe00863b7\") " pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:40 crc kubenswrapper[4646]: I1203 10:59:40.019002 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c77a1b29-f590-4342-a62a-1b3fe00863b7-config\") pod \"controller-manager-79f967984f-csln4\" (UID: \"c77a1b29-f590-4342-a62a-1b3fe00863b7\") " pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:40 crc kubenswrapper[4646]: I1203 10:59:40.023054 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c77a1b29-f590-4342-a62a-1b3fe00863b7-serving-cert\") pod \"controller-manager-79f967984f-csln4\" (UID: \"c77a1b29-f590-4342-a62a-1b3fe00863b7\") " pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:40 crc kubenswrapper[4646]: I1203 10:59:40.052480 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph5s9\" (UniqueName: \"kubernetes.io/projected/c77a1b29-f590-4342-a62a-1b3fe00863b7-kube-api-access-ph5s9\") pod \"controller-manager-79f967984f-csln4\" (UID: \"c77a1b29-f590-4342-a62a-1b3fe00863b7\") " pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:40 crc kubenswrapper[4646]: I1203 10:59:40.086149 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:40 crc kubenswrapper[4646]: I1203 10:59:40.305824 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-79f967984f-csln4"] Dec 03 10:59:40 crc kubenswrapper[4646]: I1203 10:59:40.736122 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79f967984f-csln4" event={"ID":"c77a1b29-f590-4342-a62a-1b3fe00863b7","Type":"ContainerStarted","Data":"924a6f1b23815ad9face45d8bdc119be6c1d3705e1b213927946610b91169785"} Dec 03 10:59:40 crc kubenswrapper[4646]: I1203 10:59:40.736777 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79f967984f-csln4" event={"ID":"c77a1b29-f590-4342-a62a-1b3fe00863b7","Type":"ContainerStarted","Data":"c7c96472bc8972225cae91cfb1711bb657647abcc466a8c4e1451033c50652c9"} Dec 03 10:59:41 crc kubenswrapper[4646]: I1203 10:59:41.741747 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:41 crc kubenswrapper[4646]: I1203 10:59:41.746788 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-79f967984f-csln4" Dec 03 10:59:41 crc kubenswrapper[4646]: I1203 10:59:41.770037 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-79f967984f-csln4" podStartSLOduration=4.7700209529999995 podStartE2EDuration="4.770020953s" podCreationTimestamp="2025-12-03 10:59:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:59:40.759198473 +0000 UTC m=+357.222254608" watchObservedRunningTime="2025-12-03 10:59:41.770020953 +0000 UTC m=+358.233077088" Dec 03 10:59:43 crc kubenswrapper[4646]: I1203 10:59:43.405685 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 10:59:43 crc kubenswrapper[4646]: I1203 10:59:43.406249 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 10:59:43 crc kubenswrapper[4646]: I1203 10:59:43.454816 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 10:59:43 crc kubenswrapper[4646]: I1203 10:59:43.596088 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kbtw5" Dec 03 10:59:43 crc kubenswrapper[4646]: I1203 10:59:43.596147 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kbtw5" Dec 03 10:59:43 crc kubenswrapper[4646]: I1203 10:59:43.662115 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kbtw5" Dec 03 10:59:43 crc kubenswrapper[4646]: I1203 10:59:43.789620 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kbtw5" Dec 03 10:59:43 crc kubenswrapper[4646]: I1203 10:59:43.791860 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 10:59:45 crc kubenswrapper[4646]: I1203 10:59:45.824687 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dtn6s" Dec 03 10:59:45 crc kubenswrapper[4646]: I1203 10:59:45.825129 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dtn6s" Dec 03 10:59:45 crc kubenswrapper[4646]: I1203 10:59:45.866104 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dtn6s" Dec 03 10:59:45 crc kubenswrapper[4646]: I1203 10:59:45.996845 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-82b96" Dec 03 10:59:45 crc kubenswrapper[4646]: I1203 10:59:45.996931 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-82b96" Dec 03 10:59:46 crc kubenswrapper[4646]: I1203 10:59:46.034692 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-82b96" Dec 03 10:59:46 crc kubenswrapper[4646]: I1203 10:59:46.811005 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-82b96" Dec 03 10:59:46 crc kubenswrapper[4646]: I1203 10:59:46.814562 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dtn6s" Dec 03 10:59:54 crc kubenswrapper[4646]: I1203 10:59:54.311003 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-sd54l" Dec 03 10:59:54 crc kubenswrapper[4646]: I1203 10:59:54.375740 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k4mmg"] Dec 03 10:59:55 crc kubenswrapper[4646]: I1203 10:59:55.964399 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:59:55 crc kubenswrapper[4646]: I1203 10:59:55.964712 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:59:57 crc kubenswrapper[4646]: I1203 10:59:57.868021 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh"] Dec 03 10:59:57 crc kubenswrapper[4646]: I1203 10:59:57.868280 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" podUID="ca7e2726-b8bb-45f5-af1d-e09456dfe033" containerName="route-controller-manager" containerID="cri-o://1b9fea0dc364ab19528a7a3c81f4db73d938707fff9d52c234752820f7e8af16" gracePeriod=30 Dec 03 10:59:58 crc kubenswrapper[4646]: I1203 10:59:58.831210 4646 generic.go:334] "Generic (PLEG): container finished" podID="ca7e2726-b8bb-45f5-af1d-e09456dfe033" containerID="1b9fea0dc364ab19528a7a3c81f4db73d938707fff9d52c234752820f7e8af16" exitCode=0 Dec 03 10:59:58 crc kubenswrapper[4646]: I1203 10:59:58.831365 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" event={"ID":"ca7e2726-b8bb-45f5-af1d-e09456dfe033","Type":"ContainerDied","Data":"1b9fea0dc364ab19528a7a3c81f4db73d938707fff9d52c234752820f7e8af16"} Dec 03 10:59:58 crc kubenswrapper[4646]: I1203 10:59:58.831545 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" event={"ID":"ca7e2726-b8bb-45f5-af1d-e09456dfe033","Type":"ContainerDied","Data":"a6690ddbf1910b4dd93d45a0b7967f160445fce2684f34feb63fc4581e3431c1"} Dec 03 10:59:58 crc kubenswrapper[4646]: I1203 10:59:58.831562 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6690ddbf1910b4dd93d45a0b7967f160445fce2684f34feb63fc4581e3431c1" Dec 03 10:59:58 crc kubenswrapper[4646]: I1203 10:59:58.839197 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:59:58 crc kubenswrapper[4646]: I1203 10:59:58.965569 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca7e2726-b8bb-45f5-af1d-e09456dfe033-config\") pod \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " Dec 03 10:59:58 crc kubenswrapper[4646]: I1203 10:59:58.965917 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca7e2726-b8bb-45f5-af1d-e09456dfe033-serving-cert\") pod \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " Dec 03 10:59:58 crc kubenswrapper[4646]: I1203 10:59:58.965938 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ca7e2726-b8bb-45f5-af1d-e09456dfe033-client-ca\") pod \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " Dec 03 10:59:58 crc kubenswrapper[4646]: I1203 10:59:58.965994 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvrrq\" (UniqueName: \"kubernetes.io/projected/ca7e2726-b8bb-45f5-af1d-e09456dfe033-kube-api-access-kvrrq\") pod \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\" (UID: \"ca7e2726-b8bb-45f5-af1d-e09456dfe033\") " Dec 03 10:59:58 crc kubenswrapper[4646]: I1203 10:59:58.966542 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca7e2726-b8bb-45f5-af1d-e09456dfe033-config" (OuterVolumeSpecName: "config") pod "ca7e2726-b8bb-45f5-af1d-e09456dfe033" (UID: "ca7e2726-b8bb-45f5-af1d-e09456dfe033"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:59:58 crc kubenswrapper[4646]: I1203 10:59:58.966591 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca7e2726-b8bb-45f5-af1d-e09456dfe033-client-ca" (OuterVolumeSpecName: "client-ca") pod "ca7e2726-b8bb-45f5-af1d-e09456dfe033" (UID: "ca7e2726-b8bb-45f5-af1d-e09456dfe033"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:59:58 crc kubenswrapper[4646]: I1203 10:59:58.970536 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca7e2726-b8bb-45f5-af1d-e09456dfe033-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ca7e2726-b8bb-45f5-af1d-e09456dfe033" (UID: "ca7e2726-b8bb-45f5-af1d-e09456dfe033"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:59:58 crc kubenswrapper[4646]: I1203 10:59:58.971906 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca7e2726-b8bb-45f5-af1d-e09456dfe033-kube-api-access-kvrrq" (OuterVolumeSpecName: "kube-api-access-kvrrq") pod "ca7e2726-b8bb-45f5-af1d-e09456dfe033" (UID: "ca7e2726-b8bb-45f5-af1d-e09456dfe033"). InnerVolumeSpecName "kube-api-access-kvrrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.067950 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvrrq\" (UniqueName: \"kubernetes.io/projected/ca7e2726-b8bb-45f5-af1d-e09456dfe033-kube-api-access-kvrrq\") on node \"crc\" DevicePath \"\"" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.067995 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca7e2726-b8bb-45f5-af1d-e09456dfe033-config\") on node \"crc\" DevicePath \"\"" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.068009 4646 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca7e2726-b8bb-45f5-af1d-e09456dfe033-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.068020 4646 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ca7e2726-b8bb-45f5-af1d-e09456dfe033-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.795769 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv"] Dec 03 10:59:59 crc kubenswrapper[4646]: E1203 10:59:59.796084 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca7e2726-b8bb-45f5-af1d-e09456dfe033" containerName="route-controller-manager" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.796111 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca7e2726-b8bb-45f5-af1d-e09456dfe033" containerName="route-controller-manager" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.796319 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca7e2726-b8bb-45f5-af1d-e09456dfe033" containerName="route-controller-manager" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.796974 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.819227 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv"] Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.837624 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.866138 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh"] Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.870220 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8596db5b-qtckh"] Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.877636 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e298c393-335f-4f66-b3dc-7c14e48be526-serving-cert\") pod \"route-controller-manager-67b4f8cc55-777jv\" (UID: \"e298c393-335f-4f66-b3dc-7c14e48be526\") " pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.877678 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e298c393-335f-4f66-b3dc-7c14e48be526-config\") pod \"route-controller-manager-67b4f8cc55-777jv\" (UID: \"e298c393-335f-4f66-b3dc-7c14e48be526\") " pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.877738 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbvst\" (UniqueName: \"kubernetes.io/projected/e298c393-335f-4f66-b3dc-7c14e48be526-kube-api-access-xbvst\") pod \"route-controller-manager-67b4f8cc55-777jv\" (UID: \"e298c393-335f-4f66-b3dc-7c14e48be526\") " pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.877819 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e298c393-335f-4f66-b3dc-7c14e48be526-client-ca\") pod \"route-controller-manager-67b4f8cc55-777jv\" (UID: \"e298c393-335f-4f66-b3dc-7c14e48be526\") " pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.988589 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e298c393-335f-4f66-b3dc-7c14e48be526-client-ca\") pod \"route-controller-manager-67b4f8cc55-777jv\" (UID: \"e298c393-335f-4f66-b3dc-7c14e48be526\") " pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.988662 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e298c393-335f-4f66-b3dc-7c14e48be526-serving-cert\") pod \"route-controller-manager-67b4f8cc55-777jv\" (UID: \"e298c393-335f-4f66-b3dc-7c14e48be526\") " pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.988685 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e298c393-335f-4f66-b3dc-7c14e48be526-config\") pod \"route-controller-manager-67b4f8cc55-777jv\" (UID: \"e298c393-335f-4f66-b3dc-7c14e48be526\") " pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.988730 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbvst\" (UniqueName: \"kubernetes.io/projected/e298c393-335f-4f66-b3dc-7c14e48be526-kube-api-access-xbvst\") pod \"route-controller-manager-67b4f8cc55-777jv\" (UID: \"e298c393-335f-4f66-b3dc-7c14e48be526\") " pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.989590 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e298c393-335f-4f66-b3dc-7c14e48be526-client-ca\") pod \"route-controller-manager-67b4f8cc55-777jv\" (UID: \"e298c393-335f-4f66-b3dc-7c14e48be526\") " pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.992280 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e298c393-335f-4f66-b3dc-7c14e48be526-config\") pod \"route-controller-manager-67b4f8cc55-777jv\" (UID: \"e298c393-335f-4f66-b3dc-7c14e48be526\") " pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 10:59:59 crc kubenswrapper[4646]: I1203 10:59:59.994813 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e298c393-335f-4f66-b3dc-7c14e48be526-serving-cert\") pod \"route-controller-manager-67b4f8cc55-777jv\" (UID: \"e298c393-335f-4f66-b3dc-7c14e48be526\") " pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.014142 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbvst\" (UniqueName: \"kubernetes.io/projected/e298c393-335f-4f66-b3dc-7c14e48be526-kube-api-access-xbvst\") pod \"route-controller-manager-67b4f8cc55-777jv\" (UID: \"e298c393-335f-4f66-b3dc-7c14e48be526\") " pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.117215 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.163139 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59"] Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.163963 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.166620 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.166772 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.178156 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59"] Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.192382 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/731cad85-dbe4-4118-a776-e357cc1c8a61-config-volume\") pod \"collect-profiles-29412660-w6p59\" (UID: \"731cad85-dbe4-4118-a776-e357cc1c8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.192748 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wh4p\" (UniqueName: \"kubernetes.io/projected/731cad85-dbe4-4118-a776-e357cc1c8a61-kube-api-access-4wh4p\") pod \"collect-profiles-29412660-w6p59\" (UID: \"731cad85-dbe4-4118-a776-e357cc1c8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.192778 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/731cad85-dbe4-4118-a776-e357cc1c8a61-secret-volume\") pod \"collect-profiles-29412660-w6p59\" (UID: \"731cad85-dbe4-4118-a776-e357cc1c8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.293487 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/731cad85-dbe4-4118-a776-e357cc1c8a61-config-volume\") pod \"collect-profiles-29412660-w6p59\" (UID: \"731cad85-dbe4-4118-a776-e357cc1c8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.293560 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wh4p\" (UniqueName: \"kubernetes.io/projected/731cad85-dbe4-4118-a776-e357cc1c8a61-kube-api-access-4wh4p\") pod \"collect-profiles-29412660-w6p59\" (UID: \"731cad85-dbe4-4118-a776-e357cc1c8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.293592 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/731cad85-dbe4-4118-a776-e357cc1c8a61-secret-volume\") pod \"collect-profiles-29412660-w6p59\" (UID: \"731cad85-dbe4-4118-a776-e357cc1c8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.294691 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/731cad85-dbe4-4118-a776-e357cc1c8a61-config-volume\") pod \"collect-profiles-29412660-w6p59\" (UID: \"731cad85-dbe4-4118-a776-e357cc1c8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.302179 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/731cad85-dbe4-4118-a776-e357cc1c8a61-secret-volume\") pod \"collect-profiles-29412660-w6p59\" (UID: \"731cad85-dbe4-4118-a776-e357cc1c8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.311655 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wh4p\" (UniqueName: \"kubernetes.io/projected/731cad85-dbe4-4118-a776-e357cc1c8a61-kube-api-access-4wh4p\") pod \"collect-profiles-29412660-w6p59\" (UID: \"731cad85-dbe4-4118-a776-e357cc1c8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.503547 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.566833 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv"] Dec 03 11:00:00 crc kubenswrapper[4646]: W1203 11:00:00.575550 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode298c393_335f_4f66_b3dc_7c14e48be526.slice/crio-f2c8a62f95642606346ee8f96ffc2c136ae92eb952c42a5e00e2ebc5a84967dc WatchSource:0}: Error finding container f2c8a62f95642606346ee8f96ffc2c136ae92eb952c42a5e00e2ebc5a84967dc: Status 404 returned error can't find the container with id f2c8a62f95642606346ee8f96ffc2c136ae92eb952c42a5e00e2ebc5a84967dc Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.843292 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" event={"ID":"e298c393-335f-4f66-b3dc-7c14e48be526","Type":"ContainerStarted","Data":"1e9d5cb3800efd778676a5217107668eced402bd8998bebce11618b3cb7c0558"} Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.843668 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" event={"ID":"e298c393-335f-4f66-b3dc-7c14e48be526","Type":"ContainerStarted","Data":"f2c8a62f95642606346ee8f96ffc2c136ae92eb952c42a5e00e2ebc5a84967dc"} Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.844010 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.860242 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" podStartSLOduration=3.860224869 podStartE2EDuration="3.860224869s" podCreationTimestamp="2025-12-03 10:59:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:00:00.859637062 +0000 UTC m=+377.322693227" watchObservedRunningTime="2025-12-03 11:00:00.860224869 +0000 UTC m=+377.323281004" Dec 03 11:00:00 crc kubenswrapper[4646]: I1203 11:00:00.952232 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59"] Dec 03 11:00:00 crc kubenswrapper[4646]: W1203 11:00:00.955006 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod731cad85_dbe4_4118_a776_e357cc1c8a61.slice/crio-c307655ee2906087c0b85120b45b6c8b1b5722c5368e2f02ff9648f88074acbe WatchSource:0}: Error finding container c307655ee2906087c0b85120b45b6c8b1b5722c5368e2f02ff9648f88074acbe: Status 404 returned error can't find the container with id c307655ee2906087c0b85120b45b6c8b1b5722c5368e2f02ff9648f88074acbe Dec 03 11:00:01 crc kubenswrapper[4646]: I1203 11:00:01.137699 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-67b4f8cc55-777jv" Dec 03 11:00:01 crc kubenswrapper[4646]: I1203 11:00:01.853177 4646 generic.go:334] "Generic (PLEG): container finished" podID="731cad85-dbe4-4118-a776-e357cc1c8a61" containerID="52faf881479c26858dc83a9b6578239e4735059edfdaee1e14e9a42181456cb5" exitCode=0 Dec 03 11:00:01 crc kubenswrapper[4646]: I1203 11:00:01.854193 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca7e2726-b8bb-45f5-af1d-e09456dfe033" path="/var/lib/kubelet/pods/ca7e2726-b8bb-45f5-af1d-e09456dfe033/volumes" Dec 03 11:00:01 crc kubenswrapper[4646]: I1203 11:00:01.854682 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" event={"ID":"731cad85-dbe4-4118-a776-e357cc1c8a61","Type":"ContainerDied","Data":"52faf881479c26858dc83a9b6578239e4735059edfdaee1e14e9a42181456cb5"} Dec 03 11:00:01 crc kubenswrapper[4646]: I1203 11:00:01.854718 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" event={"ID":"731cad85-dbe4-4118-a776-e357cc1c8a61","Type":"ContainerStarted","Data":"c307655ee2906087c0b85120b45b6c8b1b5722c5368e2f02ff9648f88074acbe"} Dec 03 11:00:03 crc kubenswrapper[4646]: I1203 11:00:03.126560 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" Dec 03 11:00:03 crc kubenswrapper[4646]: I1203 11:00:03.239155 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wh4p\" (UniqueName: \"kubernetes.io/projected/731cad85-dbe4-4118-a776-e357cc1c8a61-kube-api-access-4wh4p\") pod \"731cad85-dbe4-4118-a776-e357cc1c8a61\" (UID: \"731cad85-dbe4-4118-a776-e357cc1c8a61\") " Dec 03 11:00:03 crc kubenswrapper[4646]: I1203 11:00:03.239198 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/731cad85-dbe4-4118-a776-e357cc1c8a61-config-volume\") pod \"731cad85-dbe4-4118-a776-e357cc1c8a61\" (UID: \"731cad85-dbe4-4118-a776-e357cc1c8a61\") " Dec 03 11:00:03 crc kubenswrapper[4646]: I1203 11:00:03.239247 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/731cad85-dbe4-4118-a776-e357cc1c8a61-secret-volume\") pod \"731cad85-dbe4-4118-a776-e357cc1c8a61\" (UID: \"731cad85-dbe4-4118-a776-e357cc1c8a61\") " Dec 03 11:00:03 crc kubenswrapper[4646]: I1203 11:00:03.240789 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/731cad85-dbe4-4118-a776-e357cc1c8a61-config-volume" (OuterVolumeSpecName: "config-volume") pod "731cad85-dbe4-4118-a776-e357cc1c8a61" (UID: "731cad85-dbe4-4118-a776-e357cc1c8a61"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:00:03 crc kubenswrapper[4646]: I1203 11:00:03.245507 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/731cad85-dbe4-4118-a776-e357cc1c8a61-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "731cad85-dbe4-4118-a776-e357cc1c8a61" (UID: "731cad85-dbe4-4118-a776-e357cc1c8a61"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:00:03 crc kubenswrapper[4646]: I1203 11:00:03.246355 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/731cad85-dbe4-4118-a776-e357cc1c8a61-kube-api-access-4wh4p" (OuterVolumeSpecName: "kube-api-access-4wh4p") pod "731cad85-dbe4-4118-a776-e357cc1c8a61" (UID: "731cad85-dbe4-4118-a776-e357cc1c8a61"). InnerVolumeSpecName "kube-api-access-4wh4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:00:03 crc kubenswrapper[4646]: I1203 11:00:03.340159 4646 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/731cad85-dbe4-4118-a776-e357cc1c8a61-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 11:00:03 crc kubenswrapper[4646]: I1203 11:00:03.340191 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wh4p\" (UniqueName: \"kubernetes.io/projected/731cad85-dbe4-4118-a776-e357cc1c8a61-kube-api-access-4wh4p\") on node \"crc\" DevicePath \"\"" Dec 03 11:00:03 crc kubenswrapper[4646]: I1203 11:00:03.340204 4646 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/731cad85-dbe4-4118-a776-e357cc1c8a61-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 11:00:03 crc kubenswrapper[4646]: I1203 11:00:03.878115 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" event={"ID":"731cad85-dbe4-4118-a776-e357cc1c8a61","Type":"ContainerDied","Data":"c307655ee2906087c0b85120b45b6c8b1b5722c5368e2f02ff9648f88074acbe"} Dec 03 11:00:03 crc kubenswrapper[4646]: I1203 11:00:03.878246 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c307655ee2906087c0b85120b45b6c8b1b5722c5368e2f02ff9648f88074acbe" Dec 03 11:00:03 crc kubenswrapper[4646]: I1203 11:00:03.878208 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412660-w6p59" Dec 03 11:00:19 crc kubenswrapper[4646]: I1203 11:00:19.417901 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" podUID="c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd" containerName="registry" containerID="cri-o://7b0242a4a54234730cd700a00a2ca7001496a1f04f110099d7cee9038608e060" gracePeriod=30 Dec 03 11:00:19 crc kubenswrapper[4646]: I1203 11:00:19.845734 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 11:00:19 crc kubenswrapper[4646]: I1203 11:00:19.968894 4646 generic.go:334] "Generic (PLEG): container finished" podID="c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd" containerID="7b0242a4a54234730cd700a00a2ca7001496a1f04f110099d7cee9038608e060" exitCode=0 Dec 03 11:00:19 crc kubenswrapper[4646]: I1203 11:00:19.968929 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" event={"ID":"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd","Type":"ContainerDied","Data":"7b0242a4a54234730cd700a00a2ca7001496a1f04f110099d7cee9038608e060"} Dec 03 11:00:19 crc kubenswrapper[4646]: I1203 11:00:19.968953 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" event={"ID":"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd","Type":"ContainerDied","Data":"17c0251379134c085490348be935978ac89666a8e3a4371f516a4cae18e8eb4d"} Dec 03 11:00:19 crc kubenswrapper[4646]: I1203 11:00:19.968953 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k4mmg" Dec 03 11:00:19 crc kubenswrapper[4646]: I1203 11:00:19.968973 4646 scope.go:117] "RemoveContainer" containerID="7b0242a4a54234730cd700a00a2ca7001496a1f04f110099d7cee9038608e060" Dec 03 11:00:19 crc kubenswrapper[4646]: I1203 11:00:19.990305 4646 scope.go:117] "RemoveContainer" containerID="7b0242a4a54234730cd700a00a2ca7001496a1f04f110099d7cee9038608e060" Dec 03 11:00:19 crc kubenswrapper[4646]: E1203 11:00:19.990999 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b0242a4a54234730cd700a00a2ca7001496a1f04f110099d7cee9038608e060\": container with ID starting with 7b0242a4a54234730cd700a00a2ca7001496a1f04f110099d7cee9038608e060 not found: ID does not exist" containerID="7b0242a4a54234730cd700a00a2ca7001496a1f04f110099d7cee9038608e060" Dec 03 11:00:19 crc kubenswrapper[4646]: I1203 11:00:19.991044 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b0242a4a54234730cd700a00a2ca7001496a1f04f110099d7cee9038608e060"} err="failed to get container status \"7b0242a4a54234730cd700a00a2ca7001496a1f04f110099d7cee9038608e060\": rpc error: code = NotFound desc = could not find container \"7b0242a4a54234730cd700a00a2ca7001496a1f04f110099d7cee9038608e060\": container with ID starting with 7b0242a4a54234730cd700a00a2ca7001496a1f04f110099d7cee9038608e060 not found: ID does not exist" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.039392 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.039457 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tx6hp\" (UniqueName: \"kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-kube-api-access-tx6hp\") pod \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.039524 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-installation-pull-secrets\") pod \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.039566 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-ca-trust-extracted\") pod \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.039609 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-registry-tls\") pod \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.039645 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-registry-certificates\") pod \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.039684 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-trusted-ca\") pod \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.039708 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-bound-sa-token\") pod \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\" (UID: \"c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd\") " Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.040623 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.040716 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.041200 4646 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.041225 4646 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.045521 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-kube-api-access-tx6hp" (OuterVolumeSpecName: "kube-api-access-tx6hp") pod "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd"). InnerVolumeSpecName "kube-api-access-tx6hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.046212 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.047270 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.047643 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.053506 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.074607 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd" (UID: "c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.143146 4646 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.143521 4646 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.143623 4646 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.143720 4646 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.143805 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tx6hp\" (UniqueName: \"kubernetes.io/projected/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd-kube-api-access-tx6hp\") on node \"crc\" DevicePath \"\"" Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.306256 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k4mmg"] Dec 03 11:00:20 crc kubenswrapper[4646]: I1203 11:00:20.309739 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k4mmg"] Dec 03 11:00:21 crc kubenswrapper[4646]: I1203 11:00:21.857709 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd" path="/var/lib/kubelet/pods/c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd/volumes" Dec 03 11:00:25 crc kubenswrapper[4646]: I1203 11:00:25.964607 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:00:25 crc kubenswrapper[4646]: I1203 11:00:25.965044 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:00:55 crc kubenswrapper[4646]: I1203 11:00:55.964770 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:00:55 crc kubenswrapper[4646]: I1203 11:00:55.965241 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:00:55 crc kubenswrapper[4646]: I1203 11:00:55.965296 4646 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 11:00:55 crc kubenswrapper[4646]: I1203 11:00:55.965858 4646 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"18f4f8a8e5b1f28504a2c74c33b0ac967e695b60d0b4a11af354622575d39b93"} pod="openshift-machine-config-operator/machine-config-daemon-pndzt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 11:00:55 crc kubenswrapper[4646]: I1203 11:00:55.965928 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" containerID="cri-o://18f4f8a8e5b1f28504a2c74c33b0ac967e695b60d0b4a11af354622575d39b93" gracePeriod=600 Dec 03 11:00:56 crc kubenswrapper[4646]: I1203 11:00:56.165833 4646 generic.go:334] "Generic (PLEG): container finished" podID="9695d947-fc43-4625-86af-a1a3e177bb27" containerID="18f4f8a8e5b1f28504a2c74c33b0ac967e695b60d0b4a11af354622575d39b93" exitCode=0 Dec 03 11:00:56 crc kubenswrapper[4646]: I1203 11:00:56.165970 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerDied","Data":"18f4f8a8e5b1f28504a2c74c33b0ac967e695b60d0b4a11af354622575d39b93"} Dec 03 11:00:56 crc kubenswrapper[4646]: I1203 11:00:56.166088 4646 scope.go:117] "RemoveContainer" containerID="b62c42f9f6744e9654fa27b3a21a2d9ca14d81f0c3b19008ba5c09ff492c2441" Dec 03 11:00:57 crc kubenswrapper[4646]: I1203 11:00:57.177079 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerStarted","Data":"bc5af8932094babb5f9581e57117493e897a2b27daa913a38481a13fa5135b5e"} Dec 03 11:03:25 crc kubenswrapper[4646]: I1203 11:03:25.964550 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:03:25 crc kubenswrapper[4646]: I1203 11:03:25.965139 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:03:55 crc kubenswrapper[4646]: I1203 11:03:55.964112 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:03:55 crc kubenswrapper[4646]: I1203 11:03:55.964662 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.812941 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-96cjs"] Dec 03 11:04:23 crc kubenswrapper[4646]: E1203 11:04:23.813880 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="731cad85-dbe4-4118-a776-e357cc1c8a61" containerName="collect-profiles" Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.813899 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="731cad85-dbe4-4118-a776-e357cc1c8a61" containerName="collect-profiles" Dec 03 11:04:23 crc kubenswrapper[4646]: E1203 11:04:23.813937 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd" containerName="registry" Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.813947 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd" containerName="registry" Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.814157 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="731cad85-dbe4-4118-a776-e357cc1c8a61" containerName="collect-profiles" Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.814176 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7a86d4e-c1d1-40a5-a945-97a4c3ff28dd" containerName="registry" Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.822804 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-96cjs" Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.827255 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-96cjs"] Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.833155 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.833584 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xpd69"] Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.833661 4646 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-qxwqh" Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.835601 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-xpd69" Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.838751 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.840646 4646 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-kmgg9" Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.846608 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-2n8gn"] Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.847652 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-2n8gn" Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.852694 4646 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-rhtps" Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.855699 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xpd69"] Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.857776 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-2n8gn"] Dec 03 11:04:23 crc kubenswrapper[4646]: I1203 11:04:23.928167 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqljc\" (UniqueName: \"kubernetes.io/projected/345d3005-d281-404b-92bf-7f2613a262f8-kube-api-access-kqljc\") pod \"cert-manager-cainjector-7f985d654d-96cjs\" (UID: \"345d3005-d281-404b-92bf-7f2613a262f8\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-96cjs" Dec 03 11:04:24 crc kubenswrapper[4646]: I1203 11:04:24.028873 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqljc\" (UniqueName: \"kubernetes.io/projected/345d3005-d281-404b-92bf-7f2613a262f8-kube-api-access-kqljc\") pod \"cert-manager-cainjector-7f985d654d-96cjs\" (UID: \"345d3005-d281-404b-92bf-7f2613a262f8\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-96cjs" Dec 03 11:04:24 crc kubenswrapper[4646]: I1203 11:04:24.028950 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bpkj\" (UniqueName: \"kubernetes.io/projected/8a2565ee-c652-4918-8ca5-c114ce33fe17-kube-api-access-7bpkj\") pod \"cert-manager-webhook-5655c58dd6-2n8gn\" (UID: \"8a2565ee-c652-4918-8ca5-c114ce33fe17\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-2n8gn" Dec 03 11:04:24 crc kubenswrapper[4646]: I1203 11:04:24.028989 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgdsb\" (UniqueName: \"kubernetes.io/projected/22a8bf20-667c-4d61-8030-0cd69dbb0d37-kube-api-access-kgdsb\") pod \"cert-manager-5b446d88c5-xpd69\" (UID: \"22a8bf20-667c-4d61-8030-0cd69dbb0d37\") " pod="cert-manager/cert-manager-5b446d88c5-xpd69" Dec 03 11:04:24 crc kubenswrapper[4646]: I1203 11:04:24.052673 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqljc\" (UniqueName: \"kubernetes.io/projected/345d3005-d281-404b-92bf-7f2613a262f8-kube-api-access-kqljc\") pod \"cert-manager-cainjector-7f985d654d-96cjs\" (UID: \"345d3005-d281-404b-92bf-7f2613a262f8\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-96cjs" Dec 03 11:04:24 crc kubenswrapper[4646]: I1203 11:04:24.130228 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bpkj\" (UniqueName: \"kubernetes.io/projected/8a2565ee-c652-4918-8ca5-c114ce33fe17-kube-api-access-7bpkj\") pod \"cert-manager-webhook-5655c58dd6-2n8gn\" (UID: \"8a2565ee-c652-4918-8ca5-c114ce33fe17\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-2n8gn" Dec 03 11:04:24 crc kubenswrapper[4646]: I1203 11:04:24.130287 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgdsb\" (UniqueName: \"kubernetes.io/projected/22a8bf20-667c-4d61-8030-0cd69dbb0d37-kube-api-access-kgdsb\") pod \"cert-manager-5b446d88c5-xpd69\" (UID: \"22a8bf20-667c-4d61-8030-0cd69dbb0d37\") " pod="cert-manager/cert-manager-5b446d88c5-xpd69" Dec 03 11:04:24 crc kubenswrapper[4646]: I1203 11:04:24.147987 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-96cjs" Dec 03 11:04:24 crc kubenswrapper[4646]: I1203 11:04:24.159257 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bpkj\" (UniqueName: \"kubernetes.io/projected/8a2565ee-c652-4918-8ca5-c114ce33fe17-kube-api-access-7bpkj\") pod \"cert-manager-webhook-5655c58dd6-2n8gn\" (UID: \"8a2565ee-c652-4918-8ca5-c114ce33fe17\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-2n8gn" Dec 03 11:04:24 crc kubenswrapper[4646]: I1203 11:04:24.159545 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgdsb\" (UniqueName: \"kubernetes.io/projected/22a8bf20-667c-4d61-8030-0cd69dbb0d37-kube-api-access-kgdsb\") pod \"cert-manager-5b446d88c5-xpd69\" (UID: \"22a8bf20-667c-4d61-8030-0cd69dbb0d37\") " pod="cert-manager/cert-manager-5b446d88c5-xpd69" Dec 03 11:04:24 crc kubenswrapper[4646]: I1203 11:04:24.169637 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-xpd69" Dec 03 11:04:24 crc kubenswrapper[4646]: I1203 11:04:24.179677 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-2n8gn" Dec 03 11:04:24 crc kubenswrapper[4646]: I1203 11:04:24.399191 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-2n8gn"] Dec 03 11:04:24 crc kubenswrapper[4646]: I1203 11:04:24.418685 4646 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 11:04:24 crc kubenswrapper[4646]: I1203 11:04:24.668765 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-96cjs"] Dec 03 11:04:24 crc kubenswrapper[4646]: W1203 11:04:24.671426 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod345d3005_d281_404b_92bf_7f2613a262f8.slice/crio-549ca9cd1d2f1df27a2313047ca4deea66aec83bf47da3eb22bee39dff019595 WatchSource:0}: Error finding container 549ca9cd1d2f1df27a2313047ca4deea66aec83bf47da3eb22bee39dff019595: Status 404 returned error can't find the container with id 549ca9cd1d2f1df27a2313047ca4deea66aec83bf47da3eb22bee39dff019595 Dec 03 11:04:24 crc kubenswrapper[4646]: I1203 11:04:24.678718 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xpd69"] Dec 03 11:04:25 crc kubenswrapper[4646]: I1203 11:04:25.415532 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-xpd69" event={"ID":"22a8bf20-667c-4d61-8030-0cd69dbb0d37","Type":"ContainerStarted","Data":"18a04d33e730f4da190e703470ce9ba5d0d234f74a26ea7637e2e37ee4f6a37f"} Dec 03 11:04:25 crc kubenswrapper[4646]: I1203 11:04:25.416583 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-96cjs" event={"ID":"345d3005-d281-404b-92bf-7f2613a262f8","Type":"ContainerStarted","Data":"549ca9cd1d2f1df27a2313047ca4deea66aec83bf47da3eb22bee39dff019595"} Dec 03 11:04:25 crc kubenswrapper[4646]: I1203 11:04:25.417476 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-2n8gn" event={"ID":"8a2565ee-c652-4918-8ca5-c114ce33fe17","Type":"ContainerStarted","Data":"e402385c8dd9c69158ab561bfc59b302b9ecc11fcff1853bf6f5ae69afec51be"} Dec 03 11:04:25 crc kubenswrapper[4646]: I1203 11:04:25.964143 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:04:25 crc kubenswrapper[4646]: I1203 11:04:25.964238 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:04:25 crc kubenswrapper[4646]: I1203 11:04:25.964298 4646 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 11:04:25 crc kubenswrapper[4646]: I1203 11:04:25.965193 4646 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bc5af8932094babb5f9581e57117493e897a2b27daa913a38481a13fa5135b5e"} pod="openshift-machine-config-operator/machine-config-daemon-pndzt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 11:04:25 crc kubenswrapper[4646]: I1203 11:04:25.965253 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" containerID="cri-o://bc5af8932094babb5f9581e57117493e897a2b27daa913a38481a13fa5135b5e" gracePeriod=600 Dec 03 11:04:27 crc kubenswrapper[4646]: I1203 11:04:27.431952 4646 generic.go:334] "Generic (PLEG): container finished" podID="9695d947-fc43-4625-86af-a1a3e177bb27" containerID="bc5af8932094babb5f9581e57117493e897a2b27daa913a38481a13fa5135b5e" exitCode=0 Dec 03 11:04:27 crc kubenswrapper[4646]: I1203 11:04:27.432016 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerDied","Data":"bc5af8932094babb5f9581e57117493e897a2b27daa913a38481a13fa5135b5e"} Dec 03 11:04:27 crc kubenswrapper[4646]: I1203 11:04:27.432099 4646 scope.go:117] "RemoveContainer" containerID="18f4f8a8e5b1f28504a2c74c33b0ac967e695b60d0b4a11af354622575d39b93" Dec 03 11:04:28 crc kubenswrapper[4646]: I1203 11:04:28.444994 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerStarted","Data":"41b509e57c2eee648a101eefda70c1101098be6f2f4ca7c3a458c795b0cc5f55"} Dec 03 11:04:28 crc kubenswrapper[4646]: I1203 11:04:28.452764 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-2n8gn" event={"ID":"8a2565ee-c652-4918-8ca5-c114ce33fe17","Type":"ContainerStarted","Data":"7f90f2312995dcce0d14a51dcdf0c5aa9bb2f070ac1ecce908ed21cc06812920"} Dec 03 11:04:28 crc kubenswrapper[4646]: I1203 11:04:28.452997 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-2n8gn" Dec 03 11:04:28 crc kubenswrapper[4646]: I1203 11:04:28.482050 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-2n8gn" podStartSLOduration=1.872141107 podStartE2EDuration="5.482022573s" podCreationTimestamp="2025-12-03 11:04:23 +0000 UTC" firstStartedPulling="2025-12-03 11:04:24.418512011 +0000 UTC m=+640.881568146" lastFinishedPulling="2025-12-03 11:04:28.028393477 +0000 UTC m=+644.491449612" observedRunningTime="2025-12-03 11:04:28.476507478 +0000 UTC m=+644.939563633" watchObservedRunningTime="2025-12-03 11:04:28.482022573 +0000 UTC m=+644.945078708" Dec 03 11:04:29 crc kubenswrapper[4646]: I1203 11:04:29.458900 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-96cjs" event={"ID":"345d3005-d281-404b-92bf-7f2613a262f8","Type":"ContainerStarted","Data":"1b25b51a636f47b4e3bbe225878aba79cd61db2d50a5396ea0b873b7035c3a67"} Dec 03 11:04:29 crc kubenswrapper[4646]: I1203 11:04:29.461790 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-xpd69" event={"ID":"22a8bf20-667c-4d61-8030-0cd69dbb0d37","Type":"ContainerStarted","Data":"70b8d0e5e2a12cc8de32e5b8ce573552effdc140d33387951a0619b95bb2d362"} Dec 03 11:04:29 crc kubenswrapper[4646]: I1203 11:04:29.478064 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-96cjs" podStartSLOduration=2.007254651 podStartE2EDuration="6.478042846s" podCreationTimestamp="2025-12-03 11:04:23 +0000 UTC" firstStartedPulling="2025-12-03 11:04:24.675646698 +0000 UTC m=+641.138702843" lastFinishedPulling="2025-12-03 11:04:29.146434903 +0000 UTC m=+645.609491038" observedRunningTime="2025-12-03 11:04:29.474398193 +0000 UTC m=+645.937454348" watchObservedRunningTime="2025-12-03 11:04:29.478042846 +0000 UTC m=+645.941099001" Dec 03 11:04:29 crc kubenswrapper[4646]: I1203 11:04:29.497002 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-xpd69" podStartSLOduration=1.918541193 podStartE2EDuration="6.496978679s" podCreationTimestamp="2025-12-03 11:04:23 +0000 UTC" firstStartedPulling="2025-12-03 11:04:24.687978534 +0000 UTC m=+641.151034669" lastFinishedPulling="2025-12-03 11:04:29.26641602 +0000 UTC m=+645.729472155" observedRunningTime="2025-12-03 11:04:29.493766388 +0000 UTC m=+645.956822563" watchObservedRunningTime="2025-12-03 11:04:29.496978679 +0000 UTC m=+645.960034834" Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.113081 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9lfpq"] Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.113860 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovn-controller" containerID="cri-o://71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88" gracePeriod=30 Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.113967 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119" gracePeriod=30 Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.113954 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="sbdb" containerID="cri-o://ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe" gracePeriod=30 Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.113992 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovn-acl-logging" containerID="cri-o://a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca" gracePeriod=30 Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.114032 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="kube-rbac-proxy-node" containerID="cri-o://b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8" gracePeriod=30 Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.113945 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="nbdb" containerID="cri-o://7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104" gracePeriod=30 Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.113927 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="northd" containerID="cri-o://9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616" gracePeriod=30 Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.146927 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" containerID="cri-o://a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214" gracePeriod=30 Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.184746 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-2n8gn" Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.493607 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovnkube-controller/3.log" Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.495810 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovn-acl-logging/0.log" Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.496446 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovn-controller/0.log" Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.496813 4646 generic.go:334] "Generic (PLEG): container finished" podID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerID="a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214" exitCode=0 Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.496839 4646 generic.go:334] "Generic (PLEG): container finished" podID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerID="b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119" exitCode=0 Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.496851 4646 generic.go:334] "Generic (PLEG): container finished" podID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerID="b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8" exitCode=0 Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.496861 4646 generic.go:334] "Generic (PLEG): container finished" podID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerID="a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca" exitCode=143 Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.496870 4646 generic.go:334] "Generic (PLEG): container finished" podID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerID="71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88" exitCode=143 Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.496878 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerDied","Data":"a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214"} Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.496916 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerDied","Data":"b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119"} Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.496928 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerDied","Data":"b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8"} Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.496941 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerDied","Data":"a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca"} Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.496951 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerDied","Data":"71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88"} Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.496930 4646 scope.go:117] "RemoveContainer" containerID="8afba344f391525381b79ab6f9648b142cc6dafae8cf22c4e4a23804366ab034" Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.498883 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cbp2x_0b003d65-42ef-4a54-96e8-33428e114a1d/kube-multus/2.log" Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.499351 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cbp2x_0b003d65-42ef-4a54-96e8-33428e114a1d/kube-multus/1.log" Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.499404 4646 generic.go:334] "Generic (PLEG): container finished" podID="0b003d65-42ef-4a54-96e8-33428e114a1d" containerID="c364a629d49eff8128fc43ea166baeec2f873dc2d1712a4bbfc1c90a0cb2ec4c" exitCode=2 Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.499439 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cbp2x" event={"ID":"0b003d65-42ef-4a54-96e8-33428e114a1d","Type":"ContainerDied","Data":"c364a629d49eff8128fc43ea166baeec2f873dc2d1712a4bbfc1c90a0cb2ec4c"} Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.499891 4646 scope.go:117] "RemoveContainer" containerID="c364a629d49eff8128fc43ea166baeec2f873dc2d1712a4bbfc1c90a0cb2ec4c" Dec 03 11:04:34 crc kubenswrapper[4646]: E1203 11:04:34.500092 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-cbp2x_openshift-multus(0b003d65-42ef-4a54-96e8-33428e114a1d)\"" pod="openshift-multus/multus-cbp2x" podUID="0b003d65-42ef-4a54-96e8-33428e114a1d" Dec 03 11:04:34 crc kubenswrapper[4646]: I1203 11:04:34.545317 4646 scope.go:117] "RemoveContainer" containerID="ffa715e7e498fda77dee5e1e92dece871ede3ada1b73f41f8f35ff15fc8910fd" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.287448 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovn-acl-logging/0.log" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.287945 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovn-controller/0.log" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.288456 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340163 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-wzhvl"] Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.340408 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340419 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.340427 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="sbdb" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340433 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="sbdb" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.340440 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340447 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.340453 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340459 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.340469 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340475 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.340485 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovn-acl-logging" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340492 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovn-acl-logging" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.340505 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="nbdb" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340510 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="nbdb" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.340522 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="northd" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340528 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="northd" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.340538 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="kube-rbac-proxy-node" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340544 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="kube-rbac-proxy-node" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.340551 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="kubecfg-setup" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340556 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="kubecfg-setup" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.340563 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovn-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340569 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovn-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340651 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="nbdb" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340663 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340670 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="northd" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340679 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovn-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340687 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340693 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovn-acl-logging" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340701 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340708 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340713 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340722 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="sbdb" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340729 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="kube-rbac-proxy-node" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.340806 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340813 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.340895 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.340994 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.341001 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerName="ovnkube-controller" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.342563 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.367970 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-kubelet\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.368061 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-run-systemd\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.368132 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-run-netns\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.370774 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-systemd-units\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.370803 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-var-lib-openvswitch\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.370887 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-etc-openvswitch\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.370925 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-slash\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472178 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovnkube-config\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472248 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-var-lib-openvswitch\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472298 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-systemd\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472330 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-run-netns\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472387 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472422 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-ovn\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472451 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-run-ovn-kubernetes\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472496 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-env-overrides\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472553 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovn-node-metrics-cert\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472593 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-cni-bin\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472623 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472650 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-etc-openvswitch\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472698 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472700 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472698 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472694 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472750 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472730 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472698 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-systemd-units\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472777 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472814 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472839 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-node-log\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472882 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-node-log" (OuterVolumeSpecName: "node-log") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472930 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-kubelet\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472936 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472976 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-cni-netd\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.472984 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473021 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-openvswitch\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473065 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-slash\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473004 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473046 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473105 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-log-socket\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473157 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-slash" (OuterVolumeSpecName: "host-slash") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473153 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-log-socket" (OuterVolumeSpecName: "log-socket") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473214 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovnkube-script-lib\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473270 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jct4d\" (UniqueName: \"kubernetes.io/projected/8cd1981e-44c4-4251-81d1-c07c9d45258b-kube-api-access-jct4d\") pod \"8cd1981e-44c4-4251-81d1-c07c9d45258b\" (UID: \"8cd1981e-44c4-4251-81d1-c07c9d45258b\") " Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473522 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-run-ovn-kubernetes\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473572 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473588 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-log-socket\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473634 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-cni-bin\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473691 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-node-log\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473825 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-run-netns\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473884 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-run-openvswitch\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473935 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-systemd-units\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473989 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-var-lib-openvswitch\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474036 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-systemd-units\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.473836 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-run-netns\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474080 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-cni-netd\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474117 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-var-lib-openvswitch\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474141 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/09385a9b-d336-4ffc-9f76-16764e8358dd-ovn-node-metrics-cert\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474199 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/09385a9b-d336-4ffc-9f76-16764e8358dd-ovnkube-script-lib\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474267 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-etc-openvswitch\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474378 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-etc-openvswitch\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474387 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/09385a9b-d336-4ffc-9f76-16764e8358dd-ovnkube-config\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474461 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-slash\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474498 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-slash\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474518 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/09385a9b-d336-4ffc-9f76-16764e8358dd-env-overrides\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474565 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-run-ovn\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474680 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474747 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-kubelet\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474806 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-kubelet\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474819 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-run-systemd\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474886 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt2sk\" (UniqueName: \"kubernetes.io/projected/09385a9b-d336-4ffc-9f76-16764e8358dd-kube-api-access-rt2sk\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.474903 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-run-systemd\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475018 4646 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475061 4646 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475086 4646 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475111 4646 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475137 4646 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475162 4646 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475187 4646 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475209 4646 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475231 4646 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475252 4646 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475275 4646 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-node-log\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475292 4646 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475314 4646 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475447 4646 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475477 4646 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-host-slash\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475497 4646 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-log-socket\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.475513 4646 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.479914 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.481742 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cd1981e-44c4-4251-81d1-c07c9d45258b-kube-api-access-jct4d" (OuterVolumeSpecName: "kube-api-access-jct4d") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "kube-api-access-jct4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.488650 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "8cd1981e-44c4-4251-81d1-c07c9d45258b" (UID: "8cd1981e-44c4-4251-81d1-c07c9d45258b"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.514644 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovn-acl-logging/0.log" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.515885 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9lfpq_8cd1981e-44c4-4251-81d1-c07c9d45258b/ovn-controller/0.log" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.516688 4646 generic.go:334] "Generic (PLEG): container finished" podID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerID="ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe" exitCode=0 Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.516718 4646 generic.go:334] "Generic (PLEG): container finished" podID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerID="7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104" exitCode=0 Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.516728 4646 generic.go:334] "Generic (PLEG): container finished" podID="8cd1981e-44c4-4251-81d1-c07c9d45258b" containerID="9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616" exitCode=0 Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.516777 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.516793 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerDied","Data":"ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe"} Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.516823 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerDied","Data":"7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104"} Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.516838 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerDied","Data":"9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616"} Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.516850 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lfpq" event={"ID":"8cd1981e-44c4-4251-81d1-c07c9d45258b","Type":"ContainerDied","Data":"ac153287e1c2e4053fbd2a176d5e16b87ddb8ab7f8fda24eca274ecd60a99269"} Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.516868 4646 scope.go:117] "RemoveContainer" containerID="a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.518664 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cbp2x_0b003d65-42ef-4a54-96e8-33428e114a1d/kube-multus/2.log" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.533459 4646 scope.go:117] "RemoveContainer" containerID="ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.548857 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9lfpq"] Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.551454 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9lfpq"] Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.551521 4646 scope.go:117] "RemoveContainer" containerID="7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.563380 4646 scope.go:117] "RemoveContainer" containerID="9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.574568 4646 scope.go:117] "RemoveContainer" containerID="b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576132 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/09385a9b-d336-4ffc-9f76-16764e8358dd-env-overrides\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576163 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-run-ovn\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576201 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576226 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt2sk\" (UniqueName: \"kubernetes.io/projected/09385a9b-d336-4ffc-9f76-16764e8358dd-kube-api-access-rt2sk\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576245 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-run-ovn-kubernetes\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576262 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-log-socket\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576270 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-run-ovn\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576314 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-cni-bin\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576318 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576276 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-cni-bin\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576421 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-node-log\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576447 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-run-openvswitch\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576457 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-log-socket\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576496 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-cni-netd\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576537 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-cni-netd\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576560 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-node-log\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576687 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/09385a9b-d336-4ffc-9f76-16764e8358dd-env-overrides\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576718 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-run-openvswitch\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576769 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/09385a9b-d336-4ffc-9f76-16764e8358dd-ovn-node-metrics-cert\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576793 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/09385a9b-d336-4ffc-9f76-16764e8358dd-ovnkube-script-lib\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576829 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/09385a9b-d336-4ffc-9f76-16764e8358dd-ovnkube-config\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576884 4646 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8cd1981e-44c4-4251-81d1-c07c9d45258b-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576894 4646 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8cd1981e-44c4-4251-81d1-c07c9d45258b-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.576905 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jct4d\" (UniqueName: \"kubernetes.io/projected/8cd1981e-44c4-4251-81d1-c07c9d45258b-kube-api-access-jct4d\") on node \"crc\" DevicePath \"\"" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.577302 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/09385a9b-d336-4ffc-9f76-16764e8358dd-ovnkube-config\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.577764 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/09385a9b-d336-4ffc-9f76-16764e8358dd-ovnkube-script-lib\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.577829 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/09385a9b-d336-4ffc-9f76-16764e8358dd-host-run-ovn-kubernetes\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.579402 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/09385a9b-d336-4ffc-9f76-16764e8358dd-ovn-node-metrics-cert\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.587633 4646 scope.go:117] "RemoveContainer" containerID="b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.591796 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt2sk\" (UniqueName: \"kubernetes.io/projected/09385a9b-d336-4ffc-9f76-16764e8358dd-kube-api-access-rt2sk\") pod \"ovnkube-node-wzhvl\" (UID: \"09385a9b-d336-4ffc-9f76-16764e8358dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.599354 4646 scope.go:117] "RemoveContainer" containerID="a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.610633 4646 scope.go:117] "RemoveContainer" containerID="71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.623887 4646 scope.go:117] "RemoveContainer" containerID="2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.636949 4646 scope.go:117] "RemoveContainer" containerID="a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.637437 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214\": container with ID starting with a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214 not found: ID does not exist" containerID="a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.637467 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214"} err="failed to get container status \"a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214\": rpc error: code = NotFound desc = could not find container \"a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214\": container with ID starting with a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.637487 4646 scope.go:117] "RemoveContainer" containerID="ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.637841 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\": container with ID starting with ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe not found: ID does not exist" containerID="ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.637860 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe"} err="failed to get container status \"ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\": rpc error: code = NotFound desc = could not find container \"ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\": container with ID starting with ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.637872 4646 scope.go:117] "RemoveContainer" containerID="7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.638100 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\": container with ID starting with 7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104 not found: ID does not exist" containerID="7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.638122 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104"} err="failed to get container status \"7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\": rpc error: code = NotFound desc = could not find container \"7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\": container with ID starting with 7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.638133 4646 scope.go:117] "RemoveContainer" containerID="9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.638546 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\": container with ID starting with 9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616 not found: ID does not exist" containerID="9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.638591 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616"} err="failed to get container status \"9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\": rpc error: code = NotFound desc = could not find container \"9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\": container with ID starting with 9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.638616 4646 scope.go:117] "RemoveContainer" containerID="b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.639024 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\": container with ID starting with b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119 not found: ID does not exist" containerID="b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.639055 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119"} err="failed to get container status \"b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\": rpc error: code = NotFound desc = could not find container \"b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\": container with ID starting with b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.639075 4646 scope.go:117] "RemoveContainer" containerID="b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.639417 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\": container with ID starting with b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8 not found: ID does not exist" containerID="b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.639444 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8"} err="failed to get container status \"b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\": rpc error: code = NotFound desc = could not find container \"b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\": container with ID starting with b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.639465 4646 scope.go:117] "RemoveContainer" containerID="a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.639752 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\": container with ID starting with a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca not found: ID does not exist" containerID="a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.639773 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca"} err="failed to get container status \"a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\": rpc error: code = NotFound desc = could not find container \"a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\": container with ID starting with a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.639785 4646 scope.go:117] "RemoveContainer" containerID="71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.640015 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\": container with ID starting with 71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88 not found: ID does not exist" containerID="71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.640033 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88"} err="failed to get container status \"71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\": rpc error: code = NotFound desc = could not find container \"71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\": container with ID starting with 71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.640046 4646 scope.go:117] "RemoveContainer" containerID="2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8" Dec 03 11:04:35 crc kubenswrapper[4646]: E1203 11:04:35.640359 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\": container with ID starting with 2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8 not found: ID does not exist" containerID="2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.640385 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8"} err="failed to get container status \"2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\": rpc error: code = NotFound desc = could not find container \"2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\": container with ID starting with 2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.640398 4646 scope.go:117] "RemoveContainer" containerID="a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.640644 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214"} err="failed to get container status \"a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214\": rpc error: code = NotFound desc = could not find container \"a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214\": container with ID starting with a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.640672 4646 scope.go:117] "RemoveContainer" containerID="ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.640882 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe"} err="failed to get container status \"ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\": rpc error: code = NotFound desc = could not find container \"ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\": container with ID starting with ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.640903 4646 scope.go:117] "RemoveContainer" containerID="7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.641097 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104"} err="failed to get container status \"7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\": rpc error: code = NotFound desc = could not find container \"7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\": container with ID starting with 7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.641114 4646 scope.go:117] "RemoveContainer" containerID="9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.641352 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616"} err="failed to get container status \"9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\": rpc error: code = NotFound desc = could not find container \"9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\": container with ID starting with 9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.641368 4646 scope.go:117] "RemoveContainer" containerID="b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.641601 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119"} err="failed to get container status \"b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\": rpc error: code = NotFound desc = could not find container \"b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\": container with ID starting with b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.641618 4646 scope.go:117] "RemoveContainer" containerID="b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.641787 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8"} err="failed to get container status \"b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\": rpc error: code = NotFound desc = could not find container \"b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\": container with ID starting with b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.641802 4646 scope.go:117] "RemoveContainer" containerID="a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.642462 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca"} err="failed to get container status \"a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\": rpc error: code = NotFound desc = could not find container \"a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\": container with ID starting with a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.642482 4646 scope.go:117] "RemoveContainer" containerID="71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.642690 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88"} err="failed to get container status \"71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\": rpc error: code = NotFound desc = could not find container \"71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\": container with ID starting with 71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.642708 4646 scope.go:117] "RemoveContainer" containerID="2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.642929 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8"} err="failed to get container status \"2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\": rpc error: code = NotFound desc = could not find container \"2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\": container with ID starting with 2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.642956 4646 scope.go:117] "RemoveContainer" containerID="a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.643157 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214"} err="failed to get container status \"a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214\": rpc error: code = NotFound desc = could not find container \"a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214\": container with ID starting with a28cc2b842021a2569697bb94cc939270167ff23d92495f1ffe5cf499fbcc214 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.643175 4646 scope.go:117] "RemoveContainer" containerID="ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.643396 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe"} err="failed to get container status \"ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\": rpc error: code = NotFound desc = could not find container \"ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe\": container with ID starting with ec93127f71d4d7255f2cacf0b06a5c794aac5fd95389dcac9d8776c2a6874cbe not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.643418 4646 scope.go:117] "RemoveContainer" containerID="7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.643690 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104"} err="failed to get container status \"7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\": rpc error: code = NotFound desc = could not find container \"7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104\": container with ID starting with 7904bb82d467e44af3c3e9263b407bf833ce627d3f6dca950635e6a66652a104 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.643712 4646 scope.go:117] "RemoveContainer" containerID="9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.644126 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616"} err="failed to get container status \"9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\": rpc error: code = NotFound desc = could not find container \"9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616\": container with ID starting with 9e8a67ffe68b0e2539f6d697bf1f11376ee59b24c7713287a8df851f443b8616 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.644178 4646 scope.go:117] "RemoveContainer" containerID="b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.644592 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119"} err="failed to get container status \"b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\": rpc error: code = NotFound desc = could not find container \"b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119\": container with ID starting with b77c10d47641f4daba77bc1decb0e5d3ae978801f2a4e898a99970a65d0fe119 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.644613 4646 scope.go:117] "RemoveContainer" containerID="b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.644989 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8"} err="failed to get container status \"b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\": rpc error: code = NotFound desc = could not find container \"b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8\": container with ID starting with b2b043afd13c4ebab978cb9c071045acd0a5e2e2da379244e87e42489db29ec8 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.645009 4646 scope.go:117] "RemoveContainer" containerID="a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.645526 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca"} err="failed to get container status \"a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\": rpc error: code = NotFound desc = could not find container \"a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca\": container with ID starting with a61bdf643540b7cb1a8da8660ee279bcf419d227c6d40f7d709c1a39188793ca not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.645550 4646 scope.go:117] "RemoveContainer" containerID="71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.645791 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88"} err="failed to get container status \"71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\": rpc error: code = NotFound desc = could not find container \"71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88\": container with ID starting with 71b6c2d6c8eaf9e375170a3049177dfedfb235ae493d84296edb81d4422c7a88 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.645810 4646 scope.go:117] "RemoveContainer" containerID="2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.646050 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8"} err="failed to get container status \"2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\": rpc error: code = NotFound desc = could not find container \"2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8\": container with ID starting with 2db1f0e51068fef0f05cd6a08a124bbbdd3e374ac71c0ef8cde54688857c25c8 not found: ID does not exist" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.658827 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:35 crc kubenswrapper[4646]: I1203 11:04:35.856365 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cd1981e-44c4-4251-81d1-c07c9d45258b" path="/var/lib/kubelet/pods/8cd1981e-44c4-4251-81d1-c07c9d45258b/volumes" Dec 03 11:04:36 crc kubenswrapper[4646]: I1203 11:04:36.529241 4646 generic.go:334] "Generic (PLEG): container finished" podID="09385a9b-d336-4ffc-9f76-16764e8358dd" containerID="44759eddaee523411b867e59fd5f769fd0063c8c176bf8015ba47d1679d8631a" exitCode=0 Dec 03 11:04:36 crc kubenswrapper[4646]: I1203 11:04:36.529360 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" event={"ID":"09385a9b-d336-4ffc-9f76-16764e8358dd","Type":"ContainerDied","Data":"44759eddaee523411b867e59fd5f769fd0063c8c176bf8015ba47d1679d8631a"} Dec 03 11:04:36 crc kubenswrapper[4646]: I1203 11:04:36.529780 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" event={"ID":"09385a9b-d336-4ffc-9f76-16764e8358dd","Type":"ContainerStarted","Data":"ed8f373a7200a7a52f2b15384927b34453c71d37de2d28cbd42b233b311f6bd3"} Dec 03 11:04:37 crc kubenswrapper[4646]: I1203 11:04:37.535389 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" event={"ID":"09385a9b-d336-4ffc-9f76-16764e8358dd","Type":"ContainerStarted","Data":"81ad9c032572c39c23ef76dff724287627813438bb648491e3a350fc4429b66d"} Dec 03 11:04:38 crc kubenswrapper[4646]: I1203 11:04:38.545382 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" event={"ID":"09385a9b-d336-4ffc-9f76-16764e8358dd","Type":"ContainerStarted","Data":"94096b0872b2ce690563f79cab0a7d5dbc4bd17c4e08cb895ff2e51e715aebef"} Dec 03 11:04:38 crc kubenswrapper[4646]: I1203 11:04:38.545722 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" event={"ID":"09385a9b-d336-4ffc-9f76-16764e8358dd","Type":"ContainerStarted","Data":"4f462a620aa6350bcf06a2f5c8610ea0e0d4fcb3a25968600e8048d5630c4f64"} Dec 03 11:04:38 crc kubenswrapper[4646]: I1203 11:04:38.545735 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" event={"ID":"09385a9b-d336-4ffc-9f76-16764e8358dd","Type":"ContainerStarted","Data":"6f0d28933ae314c36ab2d773a3a0c9f1b60f68e80282e66f39f7b45568368163"} Dec 03 11:04:39 crc kubenswrapper[4646]: I1203 11:04:39.558242 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" event={"ID":"09385a9b-d336-4ffc-9f76-16764e8358dd","Type":"ContainerStarted","Data":"eb6dc4b350fb55374d712c7c7534f484d7c2a6347dc1c99bdcfc02f1c9db7787"} Dec 03 11:04:39 crc kubenswrapper[4646]: I1203 11:04:39.558286 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" event={"ID":"09385a9b-d336-4ffc-9f76-16764e8358dd","Type":"ContainerStarted","Data":"7b53143c494607d0fd8996443a102b90bb68665a63c9da28e5e2e0e76f82e1a9"} Dec 03 11:04:41 crc kubenswrapper[4646]: I1203 11:04:41.575801 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" event={"ID":"09385a9b-d336-4ffc-9f76-16764e8358dd","Type":"ContainerStarted","Data":"2fb58d71b8ea06b22ebbd2f50be6d12efc6496fc63d89e3265c33198ddcc198b"} Dec 03 11:04:43 crc kubenswrapper[4646]: I1203 11:04:43.595058 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" event={"ID":"09385a9b-d336-4ffc-9f76-16764e8358dd","Type":"ContainerStarted","Data":"8c8dd11515ba2536de6b2820e6ce8e35ce72d3224ec948a91e628ec9a9c0862c"} Dec 03 11:04:43 crc kubenswrapper[4646]: I1203 11:04:43.595770 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:43 crc kubenswrapper[4646]: I1203 11:04:43.595848 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:43 crc kubenswrapper[4646]: I1203 11:04:43.595864 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:43 crc kubenswrapper[4646]: I1203 11:04:43.629709 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:43 crc kubenswrapper[4646]: I1203 11:04:43.630560 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" podStartSLOduration=8.630549838 podStartE2EDuration="8.630549838s" podCreationTimestamp="2025-12-03 11:04:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:04:43.629989102 +0000 UTC m=+660.093045237" watchObservedRunningTime="2025-12-03 11:04:43.630549838 +0000 UTC m=+660.093605973" Dec 03 11:04:43 crc kubenswrapper[4646]: I1203 11:04:43.630588 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:04:46 crc kubenswrapper[4646]: I1203 11:04:46.847789 4646 scope.go:117] "RemoveContainer" containerID="c364a629d49eff8128fc43ea166baeec2f873dc2d1712a4bbfc1c90a0cb2ec4c" Dec 03 11:04:46 crc kubenswrapper[4646]: E1203 11:04:46.848717 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-cbp2x_openshift-multus(0b003d65-42ef-4a54-96e8-33428e114a1d)\"" pod="openshift-multus/multus-cbp2x" podUID="0b003d65-42ef-4a54-96e8-33428e114a1d" Dec 03 11:04:58 crc kubenswrapper[4646]: I1203 11:04:58.848467 4646 scope.go:117] "RemoveContainer" containerID="c364a629d49eff8128fc43ea166baeec2f873dc2d1712a4bbfc1c90a0cb2ec4c" Dec 03 11:05:00 crc kubenswrapper[4646]: I1203 11:05:00.702750 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cbp2x_0b003d65-42ef-4a54-96e8-33428e114a1d/kube-multus/2.log" Dec 03 11:05:00 crc kubenswrapper[4646]: I1203 11:05:00.703388 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cbp2x" event={"ID":"0b003d65-42ef-4a54-96e8-33428e114a1d","Type":"ContainerStarted","Data":"f34d189e98fbbf951b90f4a94295764f52b1a2f78b2e40b82fc3ae9c2743746e"} Dec 03 11:05:05 crc kubenswrapper[4646]: I1203 11:05:05.687167 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wzhvl" Dec 03 11:05:13 crc kubenswrapper[4646]: I1203 11:05:13.910805 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc"] Dec 03 11:05:13 crc kubenswrapper[4646]: I1203 11:05:13.912429 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" Dec 03 11:05:13 crc kubenswrapper[4646]: I1203 11:05:13.914843 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 11:05:13 crc kubenswrapper[4646]: I1203 11:05:13.925494 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc"] Dec 03 11:05:13 crc kubenswrapper[4646]: I1203 11:05:13.968881 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d37842a3-b44d-4fa4-a67a-7e1888ed8832-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc\" (UID: \"d37842a3-b44d-4fa4-a67a-7e1888ed8832\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" Dec 03 11:05:13 crc kubenswrapper[4646]: I1203 11:05:13.969145 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d37842a3-b44d-4fa4-a67a-7e1888ed8832-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc\" (UID: \"d37842a3-b44d-4fa4-a67a-7e1888ed8832\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" Dec 03 11:05:13 crc kubenswrapper[4646]: I1203 11:05:13.969246 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlfz2\" (UniqueName: \"kubernetes.io/projected/d37842a3-b44d-4fa4-a67a-7e1888ed8832-kube-api-access-mlfz2\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc\" (UID: \"d37842a3-b44d-4fa4-a67a-7e1888ed8832\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" Dec 03 11:05:14 crc kubenswrapper[4646]: I1203 11:05:14.070953 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d37842a3-b44d-4fa4-a67a-7e1888ed8832-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc\" (UID: \"d37842a3-b44d-4fa4-a67a-7e1888ed8832\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" Dec 03 11:05:14 crc kubenswrapper[4646]: I1203 11:05:14.071016 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d37842a3-b44d-4fa4-a67a-7e1888ed8832-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc\" (UID: \"d37842a3-b44d-4fa4-a67a-7e1888ed8832\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" Dec 03 11:05:14 crc kubenswrapper[4646]: I1203 11:05:14.071046 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlfz2\" (UniqueName: \"kubernetes.io/projected/d37842a3-b44d-4fa4-a67a-7e1888ed8832-kube-api-access-mlfz2\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc\" (UID: \"d37842a3-b44d-4fa4-a67a-7e1888ed8832\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" Dec 03 11:05:14 crc kubenswrapper[4646]: I1203 11:05:14.071796 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d37842a3-b44d-4fa4-a67a-7e1888ed8832-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc\" (UID: \"d37842a3-b44d-4fa4-a67a-7e1888ed8832\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" Dec 03 11:05:14 crc kubenswrapper[4646]: I1203 11:05:14.071809 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d37842a3-b44d-4fa4-a67a-7e1888ed8832-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc\" (UID: \"d37842a3-b44d-4fa4-a67a-7e1888ed8832\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" Dec 03 11:05:14 crc kubenswrapper[4646]: I1203 11:05:14.091224 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlfz2\" (UniqueName: \"kubernetes.io/projected/d37842a3-b44d-4fa4-a67a-7e1888ed8832-kube-api-access-mlfz2\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc\" (UID: \"d37842a3-b44d-4fa4-a67a-7e1888ed8832\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" Dec 03 11:05:14 crc kubenswrapper[4646]: I1203 11:05:14.230364 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" Dec 03 11:05:14 crc kubenswrapper[4646]: I1203 11:05:14.612222 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc"] Dec 03 11:05:14 crc kubenswrapper[4646]: I1203 11:05:14.791589 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" event={"ID":"d37842a3-b44d-4fa4-a67a-7e1888ed8832","Type":"ContainerStarted","Data":"06a1ecc633632b51e26f0629baac6abed65bb32684e513cd158403cec0a54f3e"} Dec 03 11:05:14 crc kubenswrapper[4646]: I1203 11:05:14.791902 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" event={"ID":"d37842a3-b44d-4fa4-a67a-7e1888ed8832","Type":"ContainerStarted","Data":"aaa81c126a5ab36a51bd17fc65f09a9bbaaece5daf7e3b4af61b8a51ce33c351"} Dec 03 11:05:15 crc kubenswrapper[4646]: I1203 11:05:15.801189 4646 generic.go:334] "Generic (PLEG): container finished" podID="d37842a3-b44d-4fa4-a67a-7e1888ed8832" containerID="06a1ecc633632b51e26f0629baac6abed65bb32684e513cd158403cec0a54f3e" exitCode=0 Dec 03 11:05:15 crc kubenswrapper[4646]: I1203 11:05:15.801238 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" event={"ID":"d37842a3-b44d-4fa4-a67a-7e1888ed8832","Type":"ContainerDied","Data":"06a1ecc633632b51e26f0629baac6abed65bb32684e513cd158403cec0a54f3e"} Dec 03 11:05:17 crc kubenswrapper[4646]: I1203 11:05:17.817936 4646 generic.go:334] "Generic (PLEG): container finished" podID="d37842a3-b44d-4fa4-a67a-7e1888ed8832" containerID="3df745818cbdc3afbf868ea57a0ea3af609b5a808c9f77f3ce6981a55b8e74c1" exitCode=0 Dec 03 11:05:17 crc kubenswrapper[4646]: I1203 11:05:17.818087 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" event={"ID":"d37842a3-b44d-4fa4-a67a-7e1888ed8832","Type":"ContainerDied","Data":"3df745818cbdc3afbf868ea57a0ea3af609b5a808c9f77f3ce6981a55b8e74c1"} Dec 03 11:05:18 crc kubenswrapper[4646]: I1203 11:05:18.828965 4646 generic.go:334] "Generic (PLEG): container finished" podID="d37842a3-b44d-4fa4-a67a-7e1888ed8832" containerID="2cadc06bd8faf0afd4db623c7b85cb72566201f13d70a373b751702722d4ccb5" exitCode=0 Dec 03 11:05:18 crc kubenswrapper[4646]: I1203 11:05:18.829024 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" event={"ID":"d37842a3-b44d-4fa4-a67a-7e1888ed8832","Type":"ContainerDied","Data":"2cadc06bd8faf0afd4db623c7b85cb72566201f13d70a373b751702722d4ccb5"} Dec 03 11:05:20 crc kubenswrapper[4646]: I1203 11:05:20.114366 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" Dec 03 11:05:20 crc kubenswrapper[4646]: I1203 11:05:20.187312 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d37842a3-b44d-4fa4-a67a-7e1888ed8832-bundle\") pod \"d37842a3-b44d-4fa4-a67a-7e1888ed8832\" (UID: \"d37842a3-b44d-4fa4-a67a-7e1888ed8832\") " Dec 03 11:05:20 crc kubenswrapper[4646]: I1203 11:05:20.187440 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlfz2\" (UniqueName: \"kubernetes.io/projected/d37842a3-b44d-4fa4-a67a-7e1888ed8832-kube-api-access-mlfz2\") pod \"d37842a3-b44d-4fa4-a67a-7e1888ed8832\" (UID: \"d37842a3-b44d-4fa4-a67a-7e1888ed8832\") " Dec 03 11:05:20 crc kubenswrapper[4646]: I1203 11:05:20.187514 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d37842a3-b44d-4fa4-a67a-7e1888ed8832-util\") pod \"d37842a3-b44d-4fa4-a67a-7e1888ed8832\" (UID: \"d37842a3-b44d-4fa4-a67a-7e1888ed8832\") " Dec 03 11:05:20 crc kubenswrapper[4646]: I1203 11:05:20.188133 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d37842a3-b44d-4fa4-a67a-7e1888ed8832-bundle" (OuterVolumeSpecName: "bundle") pod "d37842a3-b44d-4fa4-a67a-7e1888ed8832" (UID: "d37842a3-b44d-4fa4-a67a-7e1888ed8832"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:05:20 crc kubenswrapper[4646]: I1203 11:05:20.192960 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d37842a3-b44d-4fa4-a67a-7e1888ed8832-kube-api-access-mlfz2" (OuterVolumeSpecName: "kube-api-access-mlfz2") pod "d37842a3-b44d-4fa4-a67a-7e1888ed8832" (UID: "d37842a3-b44d-4fa4-a67a-7e1888ed8832"). InnerVolumeSpecName "kube-api-access-mlfz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:05:20 crc kubenswrapper[4646]: I1203 11:05:20.288552 4646 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d37842a3-b44d-4fa4-a67a-7e1888ed8832-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:05:20 crc kubenswrapper[4646]: I1203 11:05:20.288587 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlfz2\" (UniqueName: \"kubernetes.io/projected/d37842a3-b44d-4fa4-a67a-7e1888ed8832-kube-api-access-mlfz2\") on node \"crc\" DevicePath \"\"" Dec 03 11:05:20 crc kubenswrapper[4646]: I1203 11:05:20.512678 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d37842a3-b44d-4fa4-a67a-7e1888ed8832-util" (OuterVolumeSpecName: "util") pod "d37842a3-b44d-4fa4-a67a-7e1888ed8832" (UID: "d37842a3-b44d-4fa4-a67a-7e1888ed8832"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:05:20 crc kubenswrapper[4646]: I1203 11:05:20.592129 4646 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d37842a3-b44d-4fa4-a67a-7e1888ed8832-util\") on node \"crc\" DevicePath \"\"" Dec 03 11:05:20 crc kubenswrapper[4646]: I1203 11:05:20.845242 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" event={"ID":"d37842a3-b44d-4fa4-a67a-7e1888ed8832","Type":"ContainerDied","Data":"aaa81c126a5ab36a51bd17fc65f09a9bbaaece5daf7e3b4af61b8a51ce33c351"} Dec 03 11:05:20 crc kubenswrapper[4646]: I1203 11:05:20.845295 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aaa81c126a5ab36a51bd17fc65f09a9bbaaece5daf7e3b4af61b8a51ce33c351" Dec 03 11:05:20 crc kubenswrapper[4646]: I1203 11:05:20.845392 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc" Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.228804 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-5pn6t"] Dec 03 11:05:25 crc kubenswrapper[4646]: E1203 11:05:25.229304 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d37842a3-b44d-4fa4-a67a-7e1888ed8832" containerName="extract" Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.229317 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d37842a3-b44d-4fa4-a67a-7e1888ed8832" containerName="extract" Dec 03 11:05:25 crc kubenswrapper[4646]: E1203 11:05:25.229349 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d37842a3-b44d-4fa4-a67a-7e1888ed8832" containerName="pull" Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.229355 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d37842a3-b44d-4fa4-a67a-7e1888ed8832" containerName="pull" Dec 03 11:05:25 crc kubenswrapper[4646]: E1203 11:05:25.229368 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d37842a3-b44d-4fa4-a67a-7e1888ed8832" containerName="util" Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.229374 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d37842a3-b44d-4fa4-a67a-7e1888ed8832" containerName="util" Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.229461 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="d37842a3-b44d-4fa4-a67a-7e1888ed8832" containerName="extract" Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.229835 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5pn6t" Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.231620 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.231629 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.234531 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-9xk6m" Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.247648 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-5pn6t"] Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.373824 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzxgb\" (UniqueName: \"kubernetes.io/projected/9094b6b6-398b-4d47-96c5-45c236227b1c-kube-api-access-gzxgb\") pod \"nmstate-operator-5b5b58f5c8-5pn6t\" (UID: \"9094b6b6-398b-4d47-96c5-45c236227b1c\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5pn6t" Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.474999 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzxgb\" (UniqueName: \"kubernetes.io/projected/9094b6b6-398b-4d47-96c5-45c236227b1c-kube-api-access-gzxgb\") pod \"nmstate-operator-5b5b58f5c8-5pn6t\" (UID: \"9094b6b6-398b-4d47-96c5-45c236227b1c\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5pn6t" Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.496259 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzxgb\" (UniqueName: \"kubernetes.io/projected/9094b6b6-398b-4d47-96c5-45c236227b1c-kube-api-access-gzxgb\") pod \"nmstate-operator-5b5b58f5c8-5pn6t\" (UID: \"9094b6b6-398b-4d47-96c5-45c236227b1c\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5pn6t" Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.543227 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5pn6t" Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.799313 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-5pn6t"] Dec 03 11:05:25 crc kubenswrapper[4646]: I1203 11:05:25.880138 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5pn6t" event={"ID":"9094b6b6-398b-4d47-96c5-45c236227b1c","Type":"ContainerStarted","Data":"a1aef85bdb49b181a19ac519300f6518708e73caceea2c349910a05650a10676"} Dec 03 11:05:28 crc kubenswrapper[4646]: I1203 11:05:28.898837 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5pn6t" event={"ID":"9094b6b6-398b-4d47-96c5-45c236227b1c","Type":"ContainerStarted","Data":"422457436a2e665a3997e59df68681ff6ddca8c7a4935ced307e8feaf6e992c8"} Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.690699 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5pn6t" podStartSLOduration=7.059186474 podStartE2EDuration="9.690682255s" podCreationTimestamp="2025-12-03 11:05:25 +0000 UTC" firstStartedPulling="2025-12-03 11:05:25.809097742 +0000 UTC m=+702.272153877" lastFinishedPulling="2025-12-03 11:05:28.440593523 +0000 UTC m=+704.903649658" observedRunningTime="2025-12-03 11:05:28.929550163 +0000 UTC m=+705.392606298" watchObservedRunningTime="2025-12-03 11:05:34.690682255 +0000 UTC m=+711.153738390" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.691790 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-mpgkf"] Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.692577 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mpgkf" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.694608 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5k5v\" (UniqueName: \"kubernetes.io/projected/fe49c8ca-0ee9-436e-bf21-dda2e34d2032-kube-api-access-f5k5v\") pod \"nmstate-metrics-7f946cbc9-mpgkf\" (UID: \"fe49c8ca-0ee9-436e-bf21-dda2e34d2032\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mpgkf" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.694614 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-xxrw2" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.712948 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq"] Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.713978 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.722205 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-mpgkf"] Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.723195 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.793217 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq"] Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.796827 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5k5v\" (UniqueName: \"kubernetes.io/projected/fe49c8ca-0ee9-436e-bf21-dda2e34d2032-kube-api-access-f5k5v\") pod \"nmstate-metrics-7f946cbc9-mpgkf\" (UID: \"fe49c8ca-0ee9-436e-bf21-dda2e34d2032\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mpgkf" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.797025 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/1a81e55c-4071-4cd3-8e5d-5d51477c994b-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-7dflq\" (UID: \"1a81e55c-4071-4cd3-8e5d-5d51477c994b\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.797096 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq4hp\" (UniqueName: \"kubernetes.io/projected/1a81e55c-4071-4cd3-8e5d-5d51477c994b-kube-api-access-qq4hp\") pod \"nmstate-webhook-5f6d4c5ccb-7dflq\" (UID: \"1a81e55c-4071-4cd3-8e5d-5d51477c994b\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.801306 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-j8lb8"] Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.802246 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.883170 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5k5v\" (UniqueName: \"kubernetes.io/projected/fe49c8ca-0ee9-436e-bf21-dda2e34d2032-kube-api-access-f5k5v\") pod \"nmstate-metrics-7f946cbc9-mpgkf\" (UID: \"fe49c8ca-0ee9-436e-bf21-dda2e34d2032\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mpgkf" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.898216 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/2c460b1c-cc8b-4455-9ae5-ee778c542705-ovs-socket\") pod \"nmstate-handler-j8lb8\" (UID: \"2c460b1c-cc8b-4455-9ae5-ee778c542705\") " pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.898288 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/1a81e55c-4071-4cd3-8e5d-5d51477c994b-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-7dflq\" (UID: \"1a81e55c-4071-4cd3-8e5d-5d51477c994b\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.898311 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq4hp\" (UniqueName: \"kubernetes.io/projected/1a81e55c-4071-4cd3-8e5d-5d51477c994b-kube-api-access-qq4hp\") pod \"nmstate-webhook-5f6d4c5ccb-7dflq\" (UID: \"1a81e55c-4071-4cd3-8e5d-5d51477c994b\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.898350 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/2c460b1c-cc8b-4455-9ae5-ee778c542705-dbus-socket\") pod \"nmstate-handler-j8lb8\" (UID: \"2c460b1c-cc8b-4455-9ae5-ee778c542705\") " pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.898372 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85pwx\" (UniqueName: \"kubernetes.io/projected/2c460b1c-cc8b-4455-9ae5-ee778c542705-kube-api-access-85pwx\") pod \"nmstate-handler-j8lb8\" (UID: \"2c460b1c-cc8b-4455-9ae5-ee778c542705\") " pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.898401 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/2c460b1c-cc8b-4455-9ae5-ee778c542705-nmstate-lock\") pod \"nmstate-handler-j8lb8\" (UID: \"2c460b1c-cc8b-4455-9ae5-ee778c542705\") " pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:34 crc kubenswrapper[4646]: E1203 11:05:34.899029 4646 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 03 11:05:34 crc kubenswrapper[4646]: E1203 11:05:34.899094 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a81e55c-4071-4cd3-8e5d-5d51477c994b-tls-key-pair podName:1a81e55c-4071-4cd3-8e5d-5d51477c994b nodeName:}" failed. No retries permitted until 2025-12-03 11:05:35.399078011 +0000 UTC m=+711.862134146 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/1a81e55c-4071-4cd3-8e5d-5d51477c994b-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-7dflq" (UID: "1a81e55c-4071-4cd3-8e5d-5d51477c994b") : secret "openshift-nmstate-webhook" not found Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.915291 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq4hp\" (UniqueName: \"kubernetes.io/projected/1a81e55c-4071-4cd3-8e5d-5d51477c994b-kube-api-access-qq4hp\") pod \"nmstate-webhook-5f6d4c5ccb-7dflq\" (UID: \"1a81e55c-4071-4cd3-8e5d-5d51477c994b\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.931036 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d"] Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.931692 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.933789 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.934090 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.935736 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-qff48" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.941788 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d"] Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.999011 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/2c460b1c-cc8b-4455-9ae5-ee778c542705-ovs-socket\") pod \"nmstate-handler-j8lb8\" (UID: \"2c460b1c-cc8b-4455-9ae5-ee778c542705\") " pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.999263 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85bvx\" (UniqueName: \"kubernetes.io/projected/277597ef-bb30-4377-ba72-7da7a58d8d52-kube-api-access-85bvx\") pod \"nmstate-console-plugin-7fbb5f6569-nds5d\" (UID: \"277597ef-bb30-4377-ba72-7da7a58d8d52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.999158 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/2c460b1c-cc8b-4455-9ae5-ee778c542705-ovs-socket\") pod \"nmstate-handler-j8lb8\" (UID: \"2c460b1c-cc8b-4455-9ae5-ee778c542705\") " pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.999487 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/2c460b1c-cc8b-4455-9ae5-ee778c542705-dbus-socket\") pod \"nmstate-handler-j8lb8\" (UID: \"2c460b1c-cc8b-4455-9ae5-ee778c542705\") " pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.999576 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85pwx\" (UniqueName: \"kubernetes.io/projected/2c460b1c-cc8b-4455-9ae5-ee778c542705-kube-api-access-85pwx\") pod \"nmstate-handler-j8lb8\" (UID: \"2c460b1c-cc8b-4455-9ae5-ee778c542705\") " pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.999641 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/277597ef-bb30-4377-ba72-7da7a58d8d52-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-nds5d\" (UID: \"277597ef-bb30-4377-ba72-7da7a58d8d52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.999795 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/2c460b1c-cc8b-4455-9ae5-ee778c542705-dbus-socket\") pod \"nmstate-handler-j8lb8\" (UID: \"2c460b1c-cc8b-4455-9ae5-ee778c542705\") " pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.999843 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/2c460b1c-cc8b-4455-9ae5-ee778c542705-nmstate-lock\") pod \"nmstate-handler-j8lb8\" (UID: \"2c460b1c-cc8b-4455-9ae5-ee778c542705\") " pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.999899 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/2c460b1c-cc8b-4455-9ae5-ee778c542705-nmstate-lock\") pod \"nmstate-handler-j8lb8\" (UID: \"2c460b1c-cc8b-4455-9ae5-ee778c542705\") " pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:34 crc kubenswrapper[4646]: I1203 11:05:34.999910 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/277597ef-bb30-4377-ba72-7da7a58d8d52-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-nds5d\" (UID: \"277597ef-bb30-4377-ba72-7da7a58d8d52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.010198 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mpgkf" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.017392 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85pwx\" (UniqueName: \"kubernetes.io/projected/2c460b1c-cc8b-4455-9ae5-ee778c542705-kube-api-access-85pwx\") pod \"nmstate-handler-j8lb8\" (UID: \"2c460b1c-cc8b-4455-9ae5-ee778c542705\") " pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.113700 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85bvx\" (UniqueName: \"kubernetes.io/projected/277597ef-bb30-4377-ba72-7da7a58d8d52-kube-api-access-85bvx\") pod \"nmstate-console-plugin-7fbb5f6569-nds5d\" (UID: \"277597ef-bb30-4377-ba72-7da7a58d8d52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.114022 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/277597ef-bb30-4377-ba72-7da7a58d8d52-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-nds5d\" (UID: \"277597ef-bb30-4377-ba72-7da7a58d8d52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.115777 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/277597ef-bb30-4377-ba72-7da7a58d8d52-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-nds5d\" (UID: \"277597ef-bb30-4377-ba72-7da7a58d8d52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.119671 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/277597ef-bb30-4377-ba72-7da7a58d8d52-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-nds5d\" (UID: \"277597ef-bb30-4377-ba72-7da7a58d8d52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.131579 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.132875 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/277597ef-bb30-4377-ba72-7da7a58d8d52-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-nds5d\" (UID: \"277597ef-bb30-4377-ba72-7da7a58d8d52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.149733 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85bvx\" (UniqueName: \"kubernetes.io/projected/277597ef-bb30-4377-ba72-7da7a58d8d52-kube-api-access-85bvx\") pod \"nmstate-console-plugin-7fbb5f6569-nds5d\" (UID: \"277597ef-bb30-4377-ba72-7da7a58d8d52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.208320 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7cb95dcc68-8gblk"] Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.209093 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.227142 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7cb95dcc68-8gblk"] Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.260648 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.306101 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-mpgkf"] Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.319869 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-console-config\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.320057 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-console-serving-cert\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.320147 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-console-oauth-config\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.320210 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46q4v\" (UniqueName: \"kubernetes.io/projected/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-kube-api-access-46q4v\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.320271 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-oauth-serving-cert\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.320355 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-service-ca\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.320456 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-trusted-ca-bundle\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: W1203 11:05:35.357769 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe49c8ca_0ee9_436e_bf21_dda2e34d2032.slice/crio-af17dacc122d637d14f94b231b794bca21561bd0f6a563dfda9888519f718061 WatchSource:0}: Error finding container af17dacc122d637d14f94b231b794bca21561bd0f6a563dfda9888519f718061: Status 404 returned error can't find the container with id af17dacc122d637d14f94b231b794bca21561bd0f6a563dfda9888519f718061 Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.422001 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-trusted-ca-bundle\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.422055 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/1a81e55c-4071-4cd3-8e5d-5d51477c994b-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-7dflq\" (UID: \"1a81e55c-4071-4cd3-8e5d-5d51477c994b\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.422089 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-console-config\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.422110 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-console-serving-cert\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.422124 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-console-oauth-config\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.422140 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46q4v\" (UniqueName: \"kubernetes.io/projected/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-kube-api-access-46q4v\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.422156 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-oauth-serving-cert\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.422177 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-service-ca\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.423998 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-oauth-serving-cert\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.424550 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-trusted-ca-bundle\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.425233 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-service-ca\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.425654 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-console-config\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.430069 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-console-oauth-config\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.433992 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/1a81e55c-4071-4cd3-8e5d-5d51477c994b-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-7dflq\" (UID: \"1a81e55c-4071-4cd3-8e5d-5d51477c994b\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.450775 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-console-serving-cert\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.450889 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46q4v\" (UniqueName: \"kubernetes.io/projected/5d0c6f15-7068-48f8-9f4e-bd287e56a17d-kube-api-access-46q4v\") pod \"console-7cb95dcc68-8gblk\" (UID: \"5d0c6f15-7068-48f8-9f4e-bd287e56a17d\") " pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.531531 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.693594 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d"] Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.694282 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq" Dec 03 11:05:35 crc kubenswrapper[4646]: W1203 11:05:35.698656 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod277597ef_bb30_4377_ba72_7da7a58d8d52.slice/crio-52f41a1dec84f4639dff848503f8e14c82cb0adb057c27c00baed14aedb48f8c WatchSource:0}: Error finding container 52f41a1dec84f4639dff848503f8e14c82cb0adb057c27c00baed14aedb48f8c: Status 404 returned error can't find the container with id 52f41a1dec84f4639dff848503f8e14c82cb0adb057c27c00baed14aedb48f8c Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.710579 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7cb95dcc68-8gblk"] Dec 03 11:05:35 crc kubenswrapper[4646]: W1203 11:05:35.725178 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d0c6f15_7068_48f8_9f4e_bd287e56a17d.slice/crio-288e3d653f9cfc19704d8c9dbf39d3133b7f85227387d971fce6a201178c3410 WatchSource:0}: Error finding container 288e3d653f9cfc19704d8c9dbf39d3133b7f85227387d971fce6a201178c3410: Status 404 returned error can't find the container with id 288e3d653f9cfc19704d8c9dbf39d3133b7f85227387d971fce6a201178c3410 Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.884052 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq"] Dec 03 11:05:35 crc kubenswrapper[4646]: W1203 11:05:35.901141 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a81e55c_4071_4cd3_8e5d_5d51477c994b.slice/crio-a6e73aab3b08a3648eca5295772ab013e55417c34e7e1bcedb510c2e4d72b513 WatchSource:0}: Error finding container a6e73aab3b08a3648eca5295772ab013e55417c34e7e1bcedb510c2e4d72b513: Status 404 returned error can't find the container with id a6e73aab3b08a3648eca5295772ab013e55417c34e7e1bcedb510c2e4d72b513 Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.966321 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq" event={"ID":"1a81e55c-4071-4cd3-8e5d-5d51477c994b","Type":"ContainerStarted","Data":"a6e73aab3b08a3648eca5295772ab013e55417c34e7e1bcedb510c2e4d72b513"} Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.967505 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d" event={"ID":"277597ef-bb30-4377-ba72-7da7a58d8d52","Type":"ContainerStarted","Data":"52f41a1dec84f4639dff848503f8e14c82cb0adb057c27c00baed14aedb48f8c"} Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.968180 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mpgkf" event={"ID":"fe49c8ca-0ee9-436e-bf21-dda2e34d2032","Type":"ContainerStarted","Data":"af17dacc122d637d14f94b231b794bca21561bd0f6a563dfda9888519f718061"} Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.968881 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7cb95dcc68-8gblk" event={"ID":"5d0c6f15-7068-48f8-9f4e-bd287e56a17d","Type":"ContainerStarted","Data":"288e3d653f9cfc19704d8c9dbf39d3133b7f85227387d971fce6a201178c3410"} Dec 03 11:05:35 crc kubenswrapper[4646]: I1203 11:05:35.969562 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-j8lb8" event={"ID":"2c460b1c-cc8b-4455-9ae5-ee778c542705","Type":"ContainerStarted","Data":"83a1e00a9952ca6ac9e49f341455210cdd992a76d3ecaadfdb1c436bb2306a0f"} Dec 03 11:05:36 crc kubenswrapper[4646]: I1203 11:05:36.980951 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7cb95dcc68-8gblk" event={"ID":"5d0c6f15-7068-48f8-9f4e-bd287e56a17d","Type":"ContainerStarted","Data":"43128a42354f5468f21a2a51918b8abaf86580755f6ea25a20b41af539d40e75"} Dec 03 11:05:37 crc kubenswrapper[4646]: I1203 11:05:37.000577 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7cb95dcc68-8gblk" podStartSLOduration=2.000559168 podStartE2EDuration="2.000559168s" podCreationTimestamp="2025-12-03 11:05:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:05:36.99781305 +0000 UTC m=+713.460869185" watchObservedRunningTime="2025-12-03 11:05:37.000559168 +0000 UTC m=+713.463615303" Dec 03 11:05:40 crc kubenswrapper[4646]: I1203 11:05:39.999693 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-j8lb8" event={"ID":"2c460b1c-cc8b-4455-9ae5-ee778c542705","Type":"ContainerStarted","Data":"a7ad89a667271c9f8ca5d2baa9cb672612145316770f478d92d2868a6177499c"} Dec 03 11:05:40 crc kubenswrapper[4646]: I1203 11:05:40.000296 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:40 crc kubenswrapper[4646]: I1203 11:05:40.002852 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq" event={"ID":"1a81e55c-4071-4cd3-8e5d-5d51477c994b","Type":"ContainerStarted","Data":"fc5ad1c734aab331db126bb54a03979259128404c9728abdd86b3be07d3d6234"} Dec 03 11:05:40 crc kubenswrapper[4646]: I1203 11:05:40.003004 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq" Dec 03 11:05:40 crc kubenswrapper[4646]: I1203 11:05:40.005158 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d" event={"ID":"277597ef-bb30-4377-ba72-7da7a58d8d52","Type":"ContainerStarted","Data":"39a52c0690681dfe9a393991a6ba9ca6061f19fcb5c07555befd527d75a5f8bb"} Dec 03 11:05:40 crc kubenswrapper[4646]: I1203 11:05:40.007078 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mpgkf" event={"ID":"fe49c8ca-0ee9-436e-bf21-dda2e34d2032","Type":"ContainerStarted","Data":"10bc97cd66b958af36b619457c65e93e1721aaf8e8eb268133c9d0163c926265"} Dec 03 11:05:40 crc kubenswrapper[4646]: I1203 11:05:40.022574 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-j8lb8" podStartSLOduration=1.766540085 podStartE2EDuration="6.022552332s" podCreationTimestamp="2025-12-03 11:05:34 +0000 UTC" firstStartedPulling="2025-12-03 11:05:35.173149681 +0000 UTC m=+711.636205816" lastFinishedPulling="2025-12-03 11:05:39.429161928 +0000 UTC m=+715.892218063" observedRunningTime="2025-12-03 11:05:40.012240478 +0000 UTC m=+716.475296613" watchObservedRunningTime="2025-12-03 11:05:40.022552332 +0000 UTC m=+716.485608467" Dec 03 11:05:40 crc kubenswrapper[4646]: I1203 11:05:40.035116 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-nds5d" podStartSLOduration=2.324861914 podStartE2EDuration="6.03509515s" podCreationTimestamp="2025-12-03 11:05:34 +0000 UTC" firstStartedPulling="2025-12-03 11:05:35.716463462 +0000 UTC m=+712.179519597" lastFinishedPulling="2025-12-03 11:05:39.426696698 +0000 UTC m=+715.889752833" observedRunningTime="2025-12-03 11:05:40.024323183 +0000 UTC m=+716.487379318" watchObservedRunningTime="2025-12-03 11:05:40.03509515 +0000 UTC m=+716.498151295" Dec 03 11:05:40 crc kubenswrapper[4646]: I1203 11:05:40.052776 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq" podStartSLOduration=2.5071435749999997 podStartE2EDuration="6.052758463s" podCreationTimestamp="2025-12-03 11:05:34 +0000 UTC" firstStartedPulling="2025-12-03 11:05:35.903490308 +0000 UTC m=+712.366546453" lastFinishedPulling="2025-12-03 11:05:39.449105196 +0000 UTC m=+715.912161341" observedRunningTime="2025-12-03 11:05:40.051478956 +0000 UTC m=+716.514535091" watchObservedRunningTime="2025-12-03 11:05:40.052758463 +0000 UTC m=+716.515814598" Dec 03 11:05:43 crc kubenswrapper[4646]: I1203 11:05:43.025861 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mpgkf" event={"ID":"fe49c8ca-0ee9-436e-bf21-dda2e34d2032","Type":"ContainerStarted","Data":"73e37ca74cd9c872f85051c82a41218084e23607496ac05d6950ad56f2865634"} Dec 03 11:05:43 crc kubenswrapper[4646]: I1203 11:05:43.049546 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mpgkf" podStartSLOduration=2.022431764 podStartE2EDuration="9.049520909s" podCreationTimestamp="2025-12-03 11:05:34 +0000 UTC" firstStartedPulling="2025-12-03 11:05:35.359758635 +0000 UTC m=+711.822814780" lastFinishedPulling="2025-12-03 11:05:42.38684779 +0000 UTC m=+718.849903925" observedRunningTime="2025-12-03 11:05:43.045890005 +0000 UTC m=+719.508946160" watchObservedRunningTime="2025-12-03 11:05:43.049520909 +0000 UTC m=+719.512577054" Dec 03 11:05:44 crc kubenswrapper[4646]: I1203 11:05:44.048874 4646 scope.go:117] "RemoveContainer" containerID="1b9fea0dc364ab19528a7a3c81f4db73d938707fff9d52c234752820f7e8af16" Dec 03 11:05:45 crc kubenswrapper[4646]: I1203 11:05:45.167784 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-j8lb8" Dec 03 11:05:45 crc kubenswrapper[4646]: I1203 11:05:45.532383 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:45 crc kubenswrapper[4646]: I1203 11:05:45.532448 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:45 crc kubenswrapper[4646]: I1203 11:05:45.538045 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:46 crc kubenswrapper[4646]: I1203 11:05:46.053231 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7cb95dcc68-8gblk" Dec 03 11:05:46 crc kubenswrapper[4646]: I1203 11:05:46.132899 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-rvbtr"] Dec 03 11:05:55 crc kubenswrapper[4646]: I1203 11:05:55.705584 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7dflq" Dec 03 11:06:10 crc kubenswrapper[4646]: I1203 11:06:10.059705 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q"] Dec 03 11:06:10 crc kubenswrapper[4646]: I1203 11:06:10.062144 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" Dec 03 11:06:10 crc kubenswrapper[4646]: I1203 11:06:10.064272 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 11:06:10 crc kubenswrapper[4646]: I1203 11:06:10.073255 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q"] Dec 03 11:06:10 crc kubenswrapper[4646]: I1203 11:06:10.164223 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d8fd2b49-2283-4290-8e40-9d14ad64ca72-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q\" (UID: \"d8fd2b49-2283-4290-8e40-9d14ad64ca72\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" Dec 03 11:06:10 crc kubenswrapper[4646]: I1203 11:06:10.164695 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d8fd2b49-2283-4290-8e40-9d14ad64ca72-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q\" (UID: \"d8fd2b49-2283-4290-8e40-9d14ad64ca72\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" Dec 03 11:06:10 crc kubenswrapper[4646]: I1203 11:06:10.164866 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9phx\" (UniqueName: \"kubernetes.io/projected/d8fd2b49-2283-4290-8e40-9d14ad64ca72-kube-api-access-w9phx\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q\" (UID: \"d8fd2b49-2283-4290-8e40-9d14ad64ca72\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" Dec 03 11:06:10 crc kubenswrapper[4646]: I1203 11:06:10.266351 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d8fd2b49-2283-4290-8e40-9d14ad64ca72-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q\" (UID: \"d8fd2b49-2283-4290-8e40-9d14ad64ca72\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" Dec 03 11:06:10 crc kubenswrapper[4646]: I1203 11:06:10.266424 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d8fd2b49-2283-4290-8e40-9d14ad64ca72-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q\" (UID: \"d8fd2b49-2283-4290-8e40-9d14ad64ca72\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" Dec 03 11:06:10 crc kubenswrapper[4646]: I1203 11:06:10.266453 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9phx\" (UniqueName: \"kubernetes.io/projected/d8fd2b49-2283-4290-8e40-9d14ad64ca72-kube-api-access-w9phx\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q\" (UID: \"d8fd2b49-2283-4290-8e40-9d14ad64ca72\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" Dec 03 11:06:10 crc kubenswrapper[4646]: I1203 11:06:10.266884 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d8fd2b49-2283-4290-8e40-9d14ad64ca72-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q\" (UID: \"d8fd2b49-2283-4290-8e40-9d14ad64ca72\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" Dec 03 11:06:10 crc kubenswrapper[4646]: I1203 11:06:10.267188 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d8fd2b49-2283-4290-8e40-9d14ad64ca72-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q\" (UID: \"d8fd2b49-2283-4290-8e40-9d14ad64ca72\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" Dec 03 11:06:10 crc kubenswrapper[4646]: I1203 11:06:10.297377 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9phx\" (UniqueName: \"kubernetes.io/projected/d8fd2b49-2283-4290-8e40-9d14ad64ca72-kube-api-access-w9phx\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q\" (UID: \"d8fd2b49-2283-4290-8e40-9d14ad64ca72\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" Dec 03 11:06:10 crc kubenswrapper[4646]: I1203 11:06:10.445739 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" Dec 03 11:06:10 crc kubenswrapper[4646]: I1203 11:06:10.922223 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q"] Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.163374 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-rvbtr" podUID="d109aa7c-89de-44b3-852c-b3c49b2bb2a0" containerName="console" containerID="cri-o://0a2d8ecb23e9e05e7a1254d63cb75893768cf77b9b1bbe953cefbd44d09a45c9" gracePeriod=15 Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.209612 4646 generic.go:334] "Generic (PLEG): container finished" podID="d8fd2b49-2283-4290-8e40-9d14ad64ca72" containerID="9bb3768dc2e1309652962fb3ad6e2c7283fa6bdcafd7685eaa6d7ca6f3bfb867" exitCode=0 Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.209678 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" event={"ID":"d8fd2b49-2283-4290-8e40-9d14ad64ca72","Type":"ContainerDied","Data":"9bb3768dc2e1309652962fb3ad6e2c7283fa6bdcafd7685eaa6d7ca6f3bfb867"} Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.209718 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" event={"ID":"d8fd2b49-2283-4290-8e40-9d14ad64ca72","Type":"ContainerStarted","Data":"f10762a4cba1daa5b4ee1977ef6bf3c6dc631ecfc764dedd3fa66f41078c98ce"} Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.509028 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-rvbtr_d109aa7c-89de-44b3-852c-b3c49b2bb2a0/console/0.log" Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.509371 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.684497 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-oauth-serving-cert\") pod \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.684581 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-trusted-ca-bundle\") pod \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.684624 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xz4r\" (UniqueName: \"kubernetes.io/projected/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-kube-api-access-2xz4r\") pod \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.684661 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-oauth-config\") pod \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.684689 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-service-ca\") pod \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.684718 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-config\") pod \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.684742 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-serving-cert\") pod \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\" (UID: \"d109aa7c-89de-44b3-852c-b3c49b2bb2a0\") " Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.685486 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "d109aa7c-89de-44b3-852c-b3c49b2bb2a0" (UID: "d109aa7c-89de-44b3-852c-b3c49b2bb2a0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.685526 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "d109aa7c-89de-44b3-852c-b3c49b2bb2a0" (UID: "d109aa7c-89de-44b3-852c-b3c49b2bb2a0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.686144 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-config" (OuterVolumeSpecName: "console-config") pod "d109aa7c-89de-44b3-852c-b3c49b2bb2a0" (UID: "d109aa7c-89de-44b3-852c-b3c49b2bb2a0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.686510 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-service-ca" (OuterVolumeSpecName: "service-ca") pod "d109aa7c-89de-44b3-852c-b3c49b2bb2a0" (UID: "d109aa7c-89de-44b3-852c-b3c49b2bb2a0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.691110 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-kube-api-access-2xz4r" (OuterVolumeSpecName: "kube-api-access-2xz4r") pod "d109aa7c-89de-44b3-852c-b3c49b2bb2a0" (UID: "d109aa7c-89de-44b3-852c-b3c49b2bb2a0"). InnerVolumeSpecName "kube-api-access-2xz4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.695810 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "d109aa7c-89de-44b3-852c-b3c49b2bb2a0" (UID: "d109aa7c-89de-44b3-852c-b3c49b2bb2a0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.696180 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "d109aa7c-89de-44b3-852c-b3c49b2bb2a0" (UID: "d109aa7c-89de-44b3-852c-b3c49b2bb2a0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.785966 4646 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.786012 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xz4r\" (UniqueName: \"kubernetes.io/projected/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-kube-api-access-2xz4r\") on node \"crc\" DevicePath \"\"" Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.786031 4646 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.786045 4646 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.786056 4646 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.786069 4646 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 11:06:11 crc kubenswrapper[4646]: I1203 11:06:11.786080 4646 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d109aa7c-89de-44b3-852c-b3c49b2bb2a0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 11:06:12 crc kubenswrapper[4646]: I1203 11:06:12.217801 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-rvbtr_d109aa7c-89de-44b3-852c-b3c49b2bb2a0/console/0.log" Dec 03 11:06:12 crc kubenswrapper[4646]: I1203 11:06:12.217844 4646 generic.go:334] "Generic (PLEG): container finished" podID="d109aa7c-89de-44b3-852c-b3c49b2bb2a0" containerID="0a2d8ecb23e9e05e7a1254d63cb75893768cf77b9b1bbe953cefbd44d09a45c9" exitCode=2 Dec 03 11:06:12 crc kubenswrapper[4646]: I1203 11:06:12.217871 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rvbtr" event={"ID":"d109aa7c-89de-44b3-852c-b3c49b2bb2a0","Type":"ContainerDied","Data":"0a2d8ecb23e9e05e7a1254d63cb75893768cf77b9b1bbe953cefbd44d09a45c9"} Dec 03 11:06:12 crc kubenswrapper[4646]: I1203 11:06:12.217897 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rvbtr" event={"ID":"d109aa7c-89de-44b3-852c-b3c49b2bb2a0","Type":"ContainerDied","Data":"713f4c66c4c1ce1b2088c92f93b18de18ee2a87d78da93afc25d157349909056"} Dec 03 11:06:12 crc kubenswrapper[4646]: I1203 11:06:12.217913 4646 scope.go:117] "RemoveContainer" containerID="0a2d8ecb23e9e05e7a1254d63cb75893768cf77b9b1bbe953cefbd44d09a45c9" Dec 03 11:06:12 crc kubenswrapper[4646]: I1203 11:06:12.217948 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rvbtr" Dec 03 11:06:12 crc kubenswrapper[4646]: I1203 11:06:12.246627 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-rvbtr"] Dec 03 11:06:12 crc kubenswrapper[4646]: I1203 11:06:12.254790 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-rvbtr"] Dec 03 11:06:12 crc kubenswrapper[4646]: I1203 11:06:12.258291 4646 scope.go:117] "RemoveContainer" containerID="0a2d8ecb23e9e05e7a1254d63cb75893768cf77b9b1bbe953cefbd44d09a45c9" Dec 03 11:06:12 crc kubenswrapper[4646]: E1203 11:06:12.259066 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a2d8ecb23e9e05e7a1254d63cb75893768cf77b9b1bbe953cefbd44d09a45c9\": container with ID starting with 0a2d8ecb23e9e05e7a1254d63cb75893768cf77b9b1bbe953cefbd44d09a45c9 not found: ID does not exist" containerID="0a2d8ecb23e9e05e7a1254d63cb75893768cf77b9b1bbe953cefbd44d09a45c9" Dec 03 11:06:12 crc kubenswrapper[4646]: I1203 11:06:12.259118 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a2d8ecb23e9e05e7a1254d63cb75893768cf77b9b1bbe953cefbd44d09a45c9"} err="failed to get container status \"0a2d8ecb23e9e05e7a1254d63cb75893768cf77b9b1bbe953cefbd44d09a45c9\": rpc error: code = NotFound desc = could not find container \"0a2d8ecb23e9e05e7a1254d63cb75893768cf77b9b1bbe953cefbd44d09a45c9\": container with ID starting with 0a2d8ecb23e9e05e7a1254d63cb75893768cf77b9b1bbe953cefbd44d09a45c9 not found: ID does not exist" Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.224646 4646 generic.go:334] "Generic (PLEG): container finished" podID="d8fd2b49-2283-4290-8e40-9d14ad64ca72" containerID="a4eeb51ecc3c45ba95b558e41a1038e3bdfa81385467c555ff335555ce310209" exitCode=0 Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.224713 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" event={"ID":"d8fd2b49-2283-4290-8e40-9d14ad64ca72","Type":"ContainerDied","Data":"a4eeb51ecc3c45ba95b558e41a1038e3bdfa81385467c555ff335555ce310209"} Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.606368 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jz8h6"] Dec 03 11:06:13 crc kubenswrapper[4646]: E1203 11:06:13.606929 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d109aa7c-89de-44b3-852c-b3c49b2bb2a0" containerName="console" Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.606944 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d109aa7c-89de-44b3-852c-b3c49b2bb2a0" containerName="console" Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.607076 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="d109aa7c-89de-44b3-852c-b3c49b2bb2a0" containerName="console" Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.607965 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.613668 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jz8h6"] Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.746643 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-utilities\") pod \"redhat-operators-jz8h6\" (UID: \"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62\") " pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.746705 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcx42\" (UniqueName: \"kubernetes.io/projected/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-kube-api-access-hcx42\") pod \"redhat-operators-jz8h6\" (UID: \"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62\") " pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.746729 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-catalog-content\") pod \"redhat-operators-jz8h6\" (UID: \"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62\") " pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.848645 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcx42\" (UniqueName: \"kubernetes.io/projected/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-kube-api-access-hcx42\") pod \"redhat-operators-jz8h6\" (UID: \"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62\") " pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.848700 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-catalog-content\") pod \"redhat-operators-jz8h6\" (UID: \"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62\") " pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.848771 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-utilities\") pod \"redhat-operators-jz8h6\" (UID: \"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62\") " pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.849290 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-utilities\") pod \"redhat-operators-jz8h6\" (UID: \"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62\") " pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.849387 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-catalog-content\") pod \"redhat-operators-jz8h6\" (UID: \"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62\") " pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.859508 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d109aa7c-89de-44b3-852c-b3c49b2bb2a0" path="/var/lib/kubelet/pods/d109aa7c-89de-44b3-852c-b3c49b2bb2a0/volumes" Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.880662 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcx42\" (UniqueName: \"kubernetes.io/projected/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-kube-api-access-hcx42\") pod \"redhat-operators-jz8h6\" (UID: \"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62\") " pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:13 crc kubenswrapper[4646]: I1203 11:06:13.930797 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:14 crc kubenswrapper[4646]: I1203 11:06:14.140242 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jz8h6"] Dec 03 11:06:14 crc kubenswrapper[4646]: I1203 11:06:14.237471 4646 generic.go:334] "Generic (PLEG): container finished" podID="d8fd2b49-2283-4290-8e40-9d14ad64ca72" containerID="bffd1bc59be5efa4e36a697a2d501730174785a042df4773dd51d959fc6f04eb" exitCode=0 Dec 03 11:06:14 crc kubenswrapper[4646]: I1203 11:06:14.237544 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" event={"ID":"d8fd2b49-2283-4290-8e40-9d14ad64ca72","Type":"ContainerDied","Data":"bffd1bc59be5efa4e36a697a2d501730174785a042df4773dd51d959fc6f04eb"} Dec 03 11:06:14 crc kubenswrapper[4646]: I1203 11:06:14.238714 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jz8h6" event={"ID":"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62","Type":"ContainerStarted","Data":"eb42499c9812be27086df361d930fbf5ff21c9959f2a4ae3cf83cfb190bcee46"} Dec 03 11:06:15 crc kubenswrapper[4646]: I1203 11:06:15.248302 4646 generic.go:334] "Generic (PLEG): container finished" podID="0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" containerID="61f3c8d2ea56c0078b397575265a6792590c96e65d1f6944a7dd1fc09da8013e" exitCode=0 Dec 03 11:06:15 crc kubenswrapper[4646]: I1203 11:06:15.248389 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jz8h6" event={"ID":"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62","Type":"ContainerDied","Data":"61f3c8d2ea56c0078b397575265a6792590c96e65d1f6944a7dd1fc09da8013e"} Dec 03 11:06:15 crc kubenswrapper[4646]: I1203 11:06:15.487740 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" Dec 03 11:06:15 crc kubenswrapper[4646]: I1203 11:06:15.669272 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d8fd2b49-2283-4290-8e40-9d14ad64ca72-bundle\") pod \"d8fd2b49-2283-4290-8e40-9d14ad64ca72\" (UID: \"d8fd2b49-2283-4290-8e40-9d14ad64ca72\") " Dec 03 11:06:15 crc kubenswrapper[4646]: I1203 11:06:15.669387 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d8fd2b49-2283-4290-8e40-9d14ad64ca72-util\") pod \"d8fd2b49-2283-4290-8e40-9d14ad64ca72\" (UID: \"d8fd2b49-2283-4290-8e40-9d14ad64ca72\") " Dec 03 11:06:15 crc kubenswrapper[4646]: I1203 11:06:15.669430 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9phx\" (UniqueName: \"kubernetes.io/projected/d8fd2b49-2283-4290-8e40-9d14ad64ca72-kube-api-access-w9phx\") pod \"d8fd2b49-2283-4290-8e40-9d14ad64ca72\" (UID: \"d8fd2b49-2283-4290-8e40-9d14ad64ca72\") " Dec 03 11:06:15 crc kubenswrapper[4646]: I1203 11:06:15.670319 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8fd2b49-2283-4290-8e40-9d14ad64ca72-bundle" (OuterVolumeSpecName: "bundle") pod "d8fd2b49-2283-4290-8e40-9d14ad64ca72" (UID: "d8fd2b49-2283-4290-8e40-9d14ad64ca72"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:06:15 crc kubenswrapper[4646]: I1203 11:06:15.684473 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8fd2b49-2283-4290-8e40-9d14ad64ca72-kube-api-access-w9phx" (OuterVolumeSpecName: "kube-api-access-w9phx") pod "d8fd2b49-2283-4290-8e40-9d14ad64ca72" (UID: "d8fd2b49-2283-4290-8e40-9d14ad64ca72"). InnerVolumeSpecName "kube-api-access-w9phx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:06:15 crc kubenswrapper[4646]: I1203 11:06:15.690190 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8fd2b49-2283-4290-8e40-9d14ad64ca72-util" (OuterVolumeSpecName: "util") pod "d8fd2b49-2283-4290-8e40-9d14ad64ca72" (UID: "d8fd2b49-2283-4290-8e40-9d14ad64ca72"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:06:15 crc kubenswrapper[4646]: I1203 11:06:15.770614 4646 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d8fd2b49-2283-4290-8e40-9d14ad64ca72-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:06:15 crc kubenswrapper[4646]: I1203 11:06:15.770684 4646 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d8fd2b49-2283-4290-8e40-9d14ad64ca72-util\") on node \"crc\" DevicePath \"\"" Dec 03 11:06:15 crc kubenswrapper[4646]: I1203 11:06:15.770698 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9phx\" (UniqueName: \"kubernetes.io/projected/d8fd2b49-2283-4290-8e40-9d14ad64ca72-kube-api-access-w9phx\") on node \"crc\" DevicePath \"\"" Dec 03 11:06:15 crc kubenswrapper[4646]: I1203 11:06:15.949828 4646 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 11:06:16 crc kubenswrapper[4646]: I1203 11:06:16.255356 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" event={"ID":"d8fd2b49-2283-4290-8e40-9d14ad64ca72","Type":"ContainerDied","Data":"f10762a4cba1daa5b4ee1977ef6bf3c6dc631ecfc764dedd3fa66f41078c98ce"} Dec 03 11:06:16 crc kubenswrapper[4646]: I1203 11:06:16.255697 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f10762a4cba1daa5b4ee1977ef6bf3c6dc631ecfc764dedd3fa66f41078c98ce" Dec 03 11:06:16 crc kubenswrapper[4646]: I1203 11:06:16.255635 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q" Dec 03 11:06:17 crc kubenswrapper[4646]: I1203 11:06:17.262421 4646 generic.go:334] "Generic (PLEG): container finished" podID="0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" containerID="bc497c17024da75c888c2dd52b94a11cab68bb475fb80376a90a99200075bf36" exitCode=0 Dec 03 11:06:17 crc kubenswrapper[4646]: I1203 11:06:17.262466 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jz8h6" event={"ID":"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62","Type":"ContainerDied","Data":"bc497c17024da75c888c2dd52b94a11cab68bb475fb80376a90a99200075bf36"} Dec 03 11:06:18 crc kubenswrapper[4646]: I1203 11:06:18.269770 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jz8h6" event={"ID":"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62","Type":"ContainerStarted","Data":"e2bf4c690cb390c34f967d20c46fff3386dabfe79f630693dd5b2baf3ba52877"} Dec 03 11:06:18 crc kubenswrapper[4646]: I1203 11:06:18.295727 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jz8h6" podStartSLOduration=2.936347296 podStartE2EDuration="5.295708761s" podCreationTimestamp="2025-12-03 11:06:13 +0000 UTC" firstStartedPulling="2025-12-03 11:06:15.251296169 +0000 UTC m=+751.714352324" lastFinishedPulling="2025-12-03 11:06:17.610657664 +0000 UTC m=+754.073713789" observedRunningTime="2025-12-03 11:06:18.291523552 +0000 UTC m=+754.754579697" watchObservedRunningTime="2025-12-03 11:06:18.295708761 +0000 UTC m=+754.758764896" Dec 03 11:06:23 crc kubenswrapper[4646]: I1203 11:06:23.931236 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:23 crc kubenswrapper[4646]: I1203 11:06:23.932324 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.034551 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jz8h6" podUID="0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" containerName="registry-server" probeResult="failure" output=< Dec 03 11:06:25 crc kubenswrapper[4646]: timeout: failed to connect service ":50051" within 1s Dec 03 11:06:25 crc kubenswrapper[4646]: > Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.475571 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-74454568d4-nvshz"] Dec 03 11:06:25 crc kubenswrapper[4646]: E1203 11:06:25.475819 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8fd2b49-2283-4290-8e40-9d14ad64ca72" containerName="util" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.475843 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8fd2b49-2283-4290-8e40-9d14ad64ca72" containerName="util" Dec 03 11:06:25 crc kubenswrapper[4646]: E1203 11:06:25.475863 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8fd2b49-2283-4290-8e40-9d14ad64ca72" containerName="pull" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.475871 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8fd2b49-2283-4290-8e40-9d14ad64ca72" containerName="pull" Dec 03 11:06:25 crc kubenswrapper[4646]: E1203 11:06:25.475887 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8fd2b49-2283-4290-8e40-9d14ad64ca72" containerName="extract" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.475895 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8fd2b49-2283-4290-8e40-9d14ad64ca72" containerName="extract" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.476014 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8fd2b49-2283-4290-8e40-9d14ad64ca72" containerName="extract" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.476522 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.484629 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.484882 4646 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-vj6h9" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.485163 4646 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.486253 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.490279 4646 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.501162 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-74454568d4-nvshz"] Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.594779 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d60bcfb-69b3-4552-9610-e4f936439116-webhook-cert\") pod \"metallb-operator-controller-manager-74454568d4-nvshz\" (UID: \"1d60bcfb-69b3-4552-9610-e4f936439116\") " pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.594916 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d60bcfb-69b3-4552-9610-e4f936439116-apiservice-cert\") pod \"metallb-operator-controller-manager-74454568d4-nvshz\" (UID: \"1d60bcfb-69b3-4552-9610-e4f936439116\") " pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.594976 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5t2q\" (UniqueName: \"kubernetes.io/projected/1d60bcfb-69b3-4552-9610-e4f936439116-kube-api-access-h5t2q\") pod \"metallb-operator-controller-manager-74454568d4-nvshz\" (UID: \"1d60bcfb-69b3-4552-9610-e4f936439116\") " pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.695769 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5t2q\" (UniqueName: \"kubernetes.io/projected/1d60bcfb-69b3-4552-9610-e4f936439116-kube-api-access-h5t2q\") pod \"metallb-operator-controller-manager-74454568d4-nvshz\" (UID: \"1d60bcfb-69b3-4552-9610-e4f936439116\") " pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.695830 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d60bcfb-69b3-4552-9610-e4f936439116-webhook-cert\") pod \"metallb-operator-controller-manager-74454568d4-nvshz\" (UID: \"1d60bcfb-69b3-4552-9610-e4f936439116\") " pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.695874 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d60bcfb-69b3-4552-9610-e4f936439116-apiservice-cert\") pod \"metallb-operator-controller-manager-74454568d4-nvshz\" (UID: \"1d60bcfb-69b3-4552-9610-e4f936439116\") " pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.704180 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d60bcfb-69b3-4552-9610-e4f936439116-apiservice-cert\") pod \"metallb-operator-controller-manager-74454568d4-nvshz\" (UID: \"1d60bcfb-69b3-4552-9610-e4f936439116\") " pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.704403 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d60bcfb-69b3-4552-9610-e4f936439116-webhook-cert\") pod \"metallb-operator-controller-manager-74454568d4-nvshz\" (UID: \"1d60bcfb-69b3-4552-9610-e4f936439116\") " pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.754695 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5t2q\" (UniqueName: \"kubernetes.io/projected/1d60bcfb-69b3-4552-9610-e4f936439116-kube-api-access-h5t2q\") pod \"metallb-operator-controller-manager-74454568d4-nvshz\" (UID: \"1d60bcfb-69b3-4552-9610-e4f936439116\") " pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" Dec 03 11:06:25 crc kubenswrapper[4646]: I1203 11:06:25.794982 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.000361 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4"] Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.001350 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.005686 4646 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-nknnm" Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.005847 4646 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.005958 4646 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.041496 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4"] Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.102613 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-74454568d4-nvshz"] Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.107980 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/70261b97-2e67-42c1-be85-44df43cf845d-webhook-cert\") pod \"metallb-operator-webhook-server-6c975599d9-tq9v4\" (UID: \"70261b97-2e67-42c1-be85-44df43cf845d\") " pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.108052 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-928n2\" (UniqueName: \"kubernetes.io/projected/70261b97-2e67-42c1-be85-44df43cf845d-kube-api-access-928n2\") pod \"metallb-operator-webhook-server-6c975599d9-tq9v4\" (UID: \"70261b97-2e67-42c1-be85-44df43cf845d\") " pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.108133 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/70261b97-2e67-42c1-be85-44df43cf845d-apiservice-cert\") pod \"metallb-operator-webhook-server-6c975599d9-tq9v4\" (UID: \"70261b97-2e67-42c1-be85-44df43cf845d\") " pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" Dec 03 11:06:26 crc kubenswrapper[4646]: W1203 11:06:26.116602 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d60bcfb_69b3_4552_9610_e4f936439116.slice/crio-0069fec9714db97eddb857bc7b5873438b6f0419ce2c2bcbd300639dac4e92ea WatchSource:0}: Error finding container 0069fec9714db97eddb857bc7b5873438b6f0419ce2c2bcbd300639dac4e92ea: Status 404 returned error can't find the container with id 0069fec9714db97eddb857bc7b5873438b6f0419ce2c2bcbd300639dac4e92ea Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.209384 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/70261b97-2e67-42c1-be85-44df43cf845d-webhook-cert\") pod \"metallb-operator-webhook-server-6c975599d9-tq9v4\" (UID: \"70261b97-2e67-42c1-be85-44df43cf845d\") " pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.209647 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-928n2\" (UniqueName: \"kubernetes.io/projected/70261b97-2e67-42c1-be85-44df43cf845d-kube-api-access-928n2\") pod \"metallb-operator-webhook-server-6c975599d9-tq9v4\" (UID: \"70261b97-2e67-42c1-be85-44df43cf845d\") " pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.209784 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/70261b97-2e67-42c1-be85-44df43cf845d-apiservice-cert\") pod \"metallb-operator-webhook-server-6c975599d9-tq9v4\" (UID: \"70261b97-2e67-42c1-be85-44df43cf845d\") " pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.213166 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/70261b97-2e67-42c1-be85-44df43cf845d-apiservice-cert\") pod \"metallb-operator-webhook-server-6c975599d9-tq9v4\" (UID: \"70261b97-2e67-42c1-be85-44df43cf845d\") " pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.213827 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/70261b97-2e67-42c1-be85-44df43cf845d-webhook-cert\") pod \"metallb-operator-webhook-server-6c975599d9-tq9v4\" (UID: \"70261b97-2e67-42c1-be85-44df43cf845d\") " pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.232442 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-928n2\" (UniqueName: \"kubernetes.io/projected/70261b97-2e67-42c1-be85-44df43cf845d-kube-api-access-928n2\") pod \"metallb-operator-webhook-server-6c975599d9-tq9v4\" (UID: \"70261b97-2e67-42c1-be85-44df43cf845d\") " pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.312894 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" event={"ID":"1d60bcfb-69b3-4552-9610-e4f936439116","Type":"ContainerStarted","Data":"0069fec9714db97eddb857bc7b5873438b6f0419ce2c2bcbd300639dac4e92ea"} Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.361668 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" Dec 03 11:06:26 crc kubenswrapper[4646]: I1203 11:06:26.846563 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4"] Dec 03 11:06:27 crc kubenswrapper[4646]: I1203 11:06:27.332895 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" event={"ID":"70261b97-2e67-42c1-be85-44df43cf845d","Type":"ContainerStarted","Data":"c16dbef068c205089119876e37974c941116f4e63b29941b756f3d484faa45a8"} Dec 03 11:06:33 crc kubenswrapper[4646]: I1203 11:06:33.971304 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:34 crc kubenswrapper[4646]: I1203 11:06:34.014445 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:34 crc kubenswrapper[4646]: I1203 11:06:34.202571 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jz8h6"] Dec 03 11:06:35 crc kubenswrapper[4646]: I1203 11:06:35.383937 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jz8h6" podUID="0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" containerName="registry-server" containerID="cri-o://e2bf4c690cb390c34f967d20c46fff3386dabfe79f630693dd5b2baf3ba52877" gracePeriod=2 Dec 03 11:06:35 crc kubenswrapper[4646]: I1203 11:06:35.384407 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" event={"ID":"1d60bcfb-69b3-4552-9610-e4f936439116","Type":"ContainerStarted","Data":"29c542adf9442bfafa9d350f003c2d46c8925b3240e900967c638c1d3f1df60a"} Dec 03 11:06:35 crc kubenswrapper[4646]: I1203 11:06:35.384798 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" Dec 03 11:06:36 crc kubenswrapper[4646]: I1203 11:06:36.402030 4646 generic.go:334] "Generic (PLEG): container finished" podID="0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" containerID="e2bf4c690cb390c34f967d20c46fff3386dabfe79f630693dd5b2baf3ba52877" exitCode=0 Dec 03 11:06:36 crc kubenswrapper[4646]: I1203 11:06:36.402109 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jz8h6" event={"ID":"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62","Type":"ContainerDied","Data":"e2bf4c690cb390c34f967d20c46fff3386dabfe79f630693dd5b2baf3ba52877"} Dec 03 11:06:39 crc kubenswrapper[4646]: I1203 11:06:39.699695 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:39 crc kubenswrapper[4646]: I1203 11:06:39.726782 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" podStartSLOduration=5.872949657 podStartE2EDuration="14.726764213s" podCreationTimestamp="2025-12-03 11:06:25 +0000 UTC" firstStartedPulling="2025-12-03 11:06:26.122016566 +0000 UTC m=+762.585072691" lastFinishedPulling="2025-12-03 11:06:34.975831112 +0000 UTC m=+771.438887247" observedRunningTime="2025-12-03 11:06:35.415676233 +0000 UTC m=+771.878732368" watchObservedRunningTime="2025-12-03 11:06:39.726764213 +0000 UTC m=+776.189820348" Dec 03 11:06:39 crc kubenswrapper[4646]: I1203 11:06:39.803730 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcx42\" (UniqueName: \"kubernetes.io/projected/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-kube-api-access-hcx42\") pod \"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62\" (UID: \"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62\") " Dec 03 11:06:39 crc kubenswrapper[4646]: I1203 11:06:39.803795 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-catalog-content\") pod \"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62\" (UID: \"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62\") " Dec 03 11:06:39 crc kubenswrapper[4646]: I1203 11:06:39.803883 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-utilities\") pod \"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62\" (UID: \"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62\") " Dec 03 11:06:39 crc kubenswrapper[4646]: I1203 11:06:39.804946 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-utilities" (OuterVolumeSpecName: "utilities") pod "0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" (UID: "0c96199e-a9d3-4d40-8f0e-3b165a9b2b62"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:06:39 crc kubenswrapper[4646]: I1203 11:06:39.819116 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-kube-api-access-hcx42" (OuterVolumeSpecName: "kube-api-access-hcx42") pod "0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" (UID: "0c96199e-a9d3-4d40-8f0e-3b165a9b2b62"). InnerVolumeSpecName "kube-api-access-hcx42". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:06:39 crc kubenswrapper[4646]: I1203 11:06:39.905324 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 11:06:39 crc kubenswrapper[4646]: I1203 11:06:39.905396 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcx42\" (UniqueName: \"kubernetes.io/projected/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-kube-api-access-hcx42\") on node \"crc\" DevicePath \"\"" Dec 03 11:06:39 crc kubenswrapper[4646]: I1203 11:06:39.924324 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" (UID: "0c96199e-a9d3-4d40-8f0e-3b165a9b2b62"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:06:40 crc kubenswrapper[4646]: I1203 11:06:40.006704 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 11:06:40 crc kubenswrapper[4646]: I1203 11:06:40.428659 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jz8h6" event={"ID":"0c96199e-a9d3-4d40-8f0e-3b165a9b2b62","Type":"ContainerDied","Data":"eb42499c9812be27086df361d930fbf5ff21c9959f2a4ae3cf83cfb190bcee46"} Dec 03 11:06:40 crc kubenswrapper[4646]: I1203 11:06:40.428984 4646 scope.go:117] "RemoveContainer" containerID="e2bf4c690cb390c34f967d20c46fff3386dabfe79f630693dd5b2baf3ba52877" Dec 03 11:06:40 crc kubenswrapper[4646]: I1203 11:06:40.428730 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jz8h6" Dec 03 11:06:40 crc kubenswrapper[4646]: I1203 11:06:40.454613 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jz8h6"] Dec 03 11:06:40 crc kubenswrapper[4646]: I1203 11:06:40.457473 4646 scope.go:117] "RemoveContainer" containerID="bc497c17024da75c888c2dd52b94a11cab68bb475fb80376a90a99200075bf36" Dec 03 11:06:40 crc kubenswrapper[4646]: I1203 11:06:40.469486 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jz8h6"] Dec 03 11:06:40 crc kubenswrapper[4646]: I1203 11:06:40.495563 4646 scope.go:117] "RemoveContainer" containerID="61f3c8d2ea56c0078b397575265a6792590c96e65d1f6944a7dd1fc09da8013e" Dec 03 11:06:41 crc kubenswrapper[4646]: I1203 11:06:41.435791 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" event={"ID":"70261b97-2e67-42c1-be85-44df43cf845d","Type":"ContainerStarted","Data":"fdfade928121387c652713f0d96474d857c8f626f57b2bb4a7cac7aef5033c59"} Dec 03 11:06:41 crc kubenswrapper[4646]: I1203 11:06:41.435879 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" Dec 03 11:06:41 crc kubenswrapper[4646]: I1203 11:06:41.479189 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" podStartSLOduration=2.903884058 podStartE2EDuration="16.479165076s" podCreationTimestamp="2025-12-03 11:06:25 +0000 UTC" firstStartedPulling="2025-12-03 11:06:26.857273822 +0000 UTC m=+763.320329957" lastFinishedPulling="2025-12-03 11:06:40.43255484 +0000 UTC m=+776.895610975" observedRunningTime="2025-12-03 11:06:41.464662973 +0000 UTC m=+777.927719118" watchObservedRunningTime="2025-12-03 11:06:41.479165076 +0000 UTC m=+777.942221211" Dec 03 11:06:41 crc kubenswrapper[4646]: I1203 11:06:41.857741 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" path="/var/lib/kubelet/pods/0c96199e-a9d3-4d40-8f0e-3b165a9b2b62/volumes" Dec 03 11:06:55 crc kubenswrapper[4646]: I1203 11:06:55.964498 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:06:55 crc kubenswrapper[4646]: I1203 11:06:55.965046 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:06:56 crc kubenswrapper[4646]: I1203 11:06:56.365780 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6c975599d9-tq9v4" Dec 03 11:07:05 crc kubenswrapper[4646]: I1203 11:07:05.797891 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-74454568d4-nvshz" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.492836 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-knlr4"] Dec 03 11:07:06 crc kubenswrapper[4646]: E1203 11:07:06.493109 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" containerName="registry-server" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.493131 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" containerName="registry-server" Dec 03 11:07:06 crc kubenswrapper[4646]: E1203 11:07:06.493152 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" containerName="extract-content" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.493161 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" containerName="extract-content" Dec 03 11:07:06 crc kubenswrapper[4646]: E1203 11:07:06.493173 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" containerName="extract-utilities" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.493181 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" containerName="extract-utilities" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.493312 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c96199e-a9d3-4d40-8f0e-3b165a9b2b62" containerName="registry-server" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.495732 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.498097 4646 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6bhts" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.498218 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.498296 4646 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.502397 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f"] Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.503058 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.506141 4646 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.521063 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f"] Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.614811 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-vxhmd"] Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.615681 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-vxhmd" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.619486 4646 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.619486 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.620753 4646 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-xpgdz" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.621347 4646 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.643508 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-jfnxz"] Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.644515 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-jfnxz" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.650149 4646 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.663438 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-jfnxz"] Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.680562 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-frr-conf\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.680616 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-reloader\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.680639 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-frr-sockets\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.680691 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-metrics-certs\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.680712 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c5af6439-97a5-487d-b4a2-0fd0660af159-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-2ff9f\" (UID: \"c5af6439-97a5-487d-b4a2-0fd0660af159\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.680733 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dvqz\" (UniqueName: \"kubernetes.io/projected/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-kube-api-access-8dvqz\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.680756 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-metrics\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.680785 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-frr-startup\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.680817 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4rgs\" (UniqueName: \"kubernetes.io/projected/c5af6439-97a5-487d-b4a2-0fd0660af159-kube-api-access-j4rgs\") pod \"frr-k8s-webhook-server-7fcb986d4-2ff9f\" (UID: \"c5af6439-97a5-487d-b4a2-0fd0660af159\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.782470 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-frr-startup\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.782523 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4rgs\" (UniqueName: \"kubernetes.io/projected/c5af6439-97a5-487d-b4a2-0fd0660af159-kube-api-access-j4rgs\") pod \"frr-k8s-webhook-server-7fcb986d4-2ff9f\" (UID: \"c5af6439-97a5-487d-b4a2-0fd0660af159\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.782552 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzzw8\" (UniqueName: \"kubernetes.io/projected/72df546c-4ffc-46d9-9f79-54bb788a08fc-kube-api-access-fzzw8\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.782571 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d6afebba-4203-45bd-831c-f9d399d27957-cert\") pod \"controller-f8648f98b-jfnxz\" (UID: \"d6afebba-4203-45bd-831c-f9d399d27957\") " pod="metallb-system/controller-f8648f98b-jfnxz" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.782594 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbkrx\" (UniqueName: \"kubernetes.io/projected/d6afebba-4203-45bd-831c-f9d399d27957-kube-api-access-pbkrx\") pod \"controller-f8648f98b-jfnxz\" (UID: \"d6afebba-4203-45bd-831c-f9d399d27957\") " pod="metallb-system/controller-f8648f98b-jfnxz" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.782613 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6afebba-4203-45bd-831c-f9d399d27957-metrics-certs\") pod \"controller-f8648f98b-jfnxz\" (UID: \"d6afebba-4203-45bd-831c-f9d399d27957\") " pod="metallb-system/controller-f8648f98b-jfnxz" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.782635 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-frr-conf\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.782656 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-reloader\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.782810 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-frr-sockets\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.782905 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-memberlist\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.783033 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-metrics-certs\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.783109 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-frr-conf\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.783117 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-metrics-certs\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.783173 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-frr-sockets\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.783174 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c5af6439-97a5-487d-b4a2-0fd0660af159-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-2ff9f\" (UID: \"c5af6439-97a5-487d-b4a2-0fd0660af159\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" Dec 03 11:07:06 crc kubenswrapper[4646]: E1203 11:07:06.783186 4646 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 03 11:07:06 crc kubenswrapper[4646]: E1203 11:07:06.783277 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-metrics-certs podName:a44776c2-244d-4ba3-b2f5-323eef5c8b9f nodeName:}" failed. No retries permitted until 2025-12-03 11:07:07.283259109 +0000 UTC m=+803.746315354 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-metrics-certs") pod "frr-k8s-knlr4" (UID: "a44776c2-244d-4ba3-b2f5-323eef5c8b9f") : secret "frr-k8s-certs-secret" not found Dec 03 11:07:06 crc kubenswrapper[4646]: E1203 11:07:06.783222 4646 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.783231 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dvqz\" (UniqueName: \"kubernetes.io/projected/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-kube-api-access-8dvqz\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: E1203 11:07:06.783466 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c5af6439-97a5-487d-b4a2-0fd0660af159-cert podName:c5af6439-97a5-487d-b4a2-0fd0660af159 nodeName:}" failed. No retries permitted until 2025-12-03 11:07:07.283441454 +0000 UTC m=+803.746497689 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c5af6439-97a5-487d-b4a2-0fd0660af159-cert") pod "frr-k8s-webhook-server-7fcb986d4-2ff9f" (UID: "c5af6439-97a5-487d-b4a2-0fd0660af159") : secret "frr-k8s-webhook-server-cert" not found Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.783494 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-metrics\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.783545 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/72df546c-4ffc-46d9-9f79-54bb788a08fc-metallb-excludel2\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.783718 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-metrics\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.783913 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-reloader\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.784225 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-frr-startup\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.801710 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4rgs\" (UniqueName: \"kubernetes.io/projected/c5af6439-97a5-487d-b4a2-0fd0660af159-kube-api-access-j4rgs\") pod \"frr-k8s-webhook-server-7fcb986d4-2ff9f\" (UID: \"c5af6439-97a5-487d-b4a2-0fd0660af159\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.819449 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dvqz\" (UniqueName: \"kubernetes.io/projected/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-kube-api-access-8dvqz\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.884588 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/72df546c-4ffc-46d9-9f79-54bb788a08fc-metallb-excludel2\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.884848 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzzw8\" (UniqueName: \"kubernetes.io/projected/72df546c-4ffc-46d9-9f79-54bb788a08fc-kube-api-access-fzzw8\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.884866 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d6afebba-4203-45bd-831c-f9d399d27957-cert\") pod \"controller-f8648f98b-jfnxz\" (UID: \"d6afebba-4203-45bd-831c-f9d399d27957\") " pod="metallb-system/controller-f8648f98b-jfnxz" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.884886 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbkrx\" (UniqueName: \"kubernetes.io/projected/d6afebba-4203-45bd-831c-f9d399d27957-kube-api-access-pbkrx\") pod \"controller-f8648f98b-jfnxz\" (UID: \"d6afebba-4203-45bd-831c-f9d399d27957\") " pod="metallb-system/controller-f8648f98b-jfnxz" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.884911 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6afebba-4203-45bd-831c-f9d399d27957-metrics-certs\") pod \"controller-f8648f98b-jfnxz\" (UID: \"d6afebba-4203-45bd-831c-f9d399d27957\") " pod="metallb-system/controller-f8648f98b-jfnxz" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.884943 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-memberlist\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.884973 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-metrics-certs\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:06 crc kubenswrapper[4646]: E1203 11:07:06.885079 4646 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 11:07:06 crc kubenswrapper[4646]: E1203 11:07:06.885145 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-memberlist podName:72df546c-4ffc-46d9-9f79-54bb788a08fc nodeName:}" failed. No retries permitted until 2025-12-03 11:07:07.385124999 +0000 UTC m=+803.848181134 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-memberlist") pod "speaker-vxhmd" (UID: "72df546c-4ffc-46d9-9f79-54bb788a08fc") : secret "metallb-memberlist" not found Dec 03 11:07:06 crc kubenswrapper[4646]: E1203 11:07:06.885211 4646 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 03 11:07:06 crc kubenswrapper[4646]: E1203 11:07:06.885313 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-metrics-certs podName:72df546c-4ffc-46d9-9f79-54bb788a08fc nodeName:}" failed. No retries permitted until 2025-12-03 11:07:07.385295053 +0000 UTC m=+803.848351188 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-metrics-certs") pod "speaker-vxhmd" (UID: "72df546c-4ffc-46d9-9f79-54bb788a08fc") : secret "speaker-certs-secret" not found Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.886004 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/72df546c-4ffc-46d9-9f79-54bb788a08fc-metallb-excludel2\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.887990 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6afebba-4203-45bd-831c-f9d399d27957-metrics-certs\") pod \"controller-f8648f98b-jfnxz\" (UID: \"d6afebba-4203-45bd-831c-f9d399d27957\") " pod="metallb-system/controller-f8648f98b-jfnxz" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.888266 4646 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.899440 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d6afebba-4203-45bd-831c-f9d399d27957-cert\") pod \"controller-f8648f98b-jfnxz\" (UID: \"d6afebba-4203-45bd-831c-f9d399d27957\") " pod="metallb-system/controller-f8648f98b-jfnxz" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.924653 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbkrx\" (UniqueName: \"kubernetes.io/projected/d6afebba-4203-45bd-831c-f9d399d27957-kube-api-access-pbkrx\") pod \"controller-f8648f98b-jfnxz\" (UID: \"d6afebba-4203-45bd-831c-f9d399d27957\") " pod="metallb-system/controller-f8648f98b-jfnxz" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.936113 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzzw8\" (UniqueName: \"kubernetes.io/projected/72df546c-4ffc-46d9-9f79-54bb788a08fc-kube-api-access-fzzw8\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:06 crc kubenswrapper[4646]: I1203 11:07:06.957345 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-jfnxz" Dec 03 11:07:07 crc kubenswrapper[4646]: I1203 11:07:07.289650 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-metrics-certs\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:07 crc kubenswrapper[4646]: I1203 11:07:07.290488 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c5af6439-97a5-487d-b4a2-0fd0660af159-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-2ff9f\" (UID: \"c5af6439-97a5-487d-b4a2-0fd0660af159\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" Dec 03 11:07:07 crc kubenswrapper[4646]: I1203 11:07:07.294030 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c5af6439-97a5-487d-b4a2-0fd0660af159-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-2ff9f\" (UID: \"c5af6439-97a5-487d-b4a2-0fd0660af159\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" Dec 03 11:07:07 crc kubenswrapper[4646]: I1203 11:07:07.294196 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a44776c2-244d-4ba3-b2f5-323eef5c8b9f-metrics-certs\") pod \"frr-k8s-knlr4\" (UID: \"a44776c2-244d-4ba3-b2f5-323eef5c8b9f\") " pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:07 crc kubenswrapper[4646]: I1203 11:07:07.384393 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-jfnxz"] Dec 03 11:07:07 crc kubenswrapper[4646]: W1203 11:07:07.390884 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6afebba_4203_45bd_831c_f9d399d27957.slice/crio-1644cfeeb3318224091fa71c7ddb7344db9dadf01fc1906947bd7fad40decd4b WatchSource:0}: Error finding container 1644cfeeb3318224091fa71c7ddb7344db9dadf01fc1906947bd7fad40decd4b: Status 404 returned error can't find the container with id 1644cfeeb3318224091fa71c7ddb7344db9dadf01fc1906947bd7fad40decd4b Dec 03 11:07:07 crc kubenswrapper[4646]: I1203 11:07:07.391387 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-memberlist\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:07 crc kubenswrapper[4646]: I1203 11:07:07.391457 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-metrics-certs\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:07 crc kubenswrapper[4646]: E1203 11:07:07.391576 4646 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 11:07:07 crc kubenswrapper[4646]: E1203 11:07:07.391638 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-memberlist podName:72df546c-4ffc-46d9-9f79-54bb788a08fc nodeName:}" failed. No retries permitted until 2025-12-03 11:07:08.391616986 +0000 UTC m=+804.854673191 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-memberlist") pod "speaker-vxhmd" (UID: "72df546c-4ffc-46d9-9f79-54bb788a08fc") : secret "metallb-memberlist" not found Dec 03 11:07:07 crc kubenswrapper[4646]: I1203 11:07:07.395529 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-metrics-certs\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:07 crc kubenswrapper[4646]: I1203 11:07:07.423102 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:07 crc kubenswrapper[4646]: I1203 11:07:07.434708 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" Dec 03 11:07:07 crc kubenswrapper[4646]: I1203 11:07:07.601195 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-jfnxz" event={"ID":"d6afebba-4203-45bd-831c-f9d399d27957","Type":"ContainerStarted","Data":"1644cfeeb3318224091fa71c7ddb7344db9dadf01fc1906947bd7fad40decd4b"} Dec 03 11:07:07 crc kubenswrapper[4646]: I1203 11:07:07.643479 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f"] Dec 03 11:07:08 crc kubenswrapper[4646]: I1203 11:07:08.403158 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-memberlist\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:08 crc kubenswrapper[4646]: E1203 11:07:08.403326 4646 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 11:07:08 crc kubenswrapper[4646]: E1203 11:07:08.403954 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-memberlist podName:72df546c-4ffc-46d9-9f79-54bb788a08fc nodeName:}" failed. No retries permitted until 2025-12-03 11:07:10.403934402 +0000 UTC m=+806.866990547 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-memberlist") pod "speaker-vxhmd" (UID: "72df546c-4ffc-46d9-9f79-54bb788a08fc") : secret "metallb-memberlist" not found Dec 03 11:07:08 crc kubenswrapper[4646]: I1203 11:07:08.607043 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" event={"ID":"c5af6439-97a5-487d-b4a2-0fd0660af159","Type":"ContainerStarted","Data":"5ae3bfc43454d45d9839d9c9de1cbc4c3547d305e7c860b84b88ca6f9a6f0a77"} Dec 03 11:07:08 crc kubenswrapper[4646]: I1203 11:07:08.607891 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-jfnxz" event={"ID":"d6afebba-4203-45bd-831c-f9d399d27957","Type":"ContainerStarted","Data":"bc257feedd9d6aca53896d08d37abe127ad2ff9be0016f2e8fc1dd7d8e33cfae"} Dec 03 11:07:08 crc kubenswrapper[4646]: I1203 11:07:08.608526 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-knlr4" event={"ID":"a44776c2-244d-4ba3-b2f5-323eef5c8b9f","Type":"ContainerStarted","Data":"433c39b3c77d56739bf721bdc34806bfa3420186ba2c6ab5e852822b6c7a84fe"} Dec 03 11:07:09 crc kubenswrapper[4646]: I1203 11:07:09.621431 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-jfnxz" event={"ID":"d6afebba-4203-45bd-831c-f9d399d27957","Type":"ContainerStarted","Data":"5753521f974847fa469e3ac13cb32638ddd28d8f81c76e2c60f336a5fccb9d8c"} Dec 03 11:07:09 crc kubenswrapper[4646]: I1203 11:07:09.621774 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-jfnxz" Dec 03 11:07:09 crc kubenswrapper[4646]: I1203 11:07:09.683724 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-jfnxz" podStartSLOduration=3.683703442 podStartE2EDuration="3.683703442s" podCreationTimestamp="2025-12-03 11:07:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:07:09.679829254 +0000 UTC m=+806.142885389" watchObservedRunningTime="2025-12-03 11:07:09.683703442 +0000 UTC m=+806.146759577" Dec 03 11:07:10 crc kubenswrapper[4646]: I1203 11:07:10.432113 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-memberlist\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:10 crc kubenswrapper[4646]: I1203 11:07:10.457260 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72df546c-4ffc-46d9-9f79-54bb788a08fc-memberlist\") pod \"speaker-vxhmd\" (UID: \"72df546c-4ffc-46d9-9f79-54bb788a08fc\") " pod="metallb-system/speaker-vxhmd" Dec 03 11:07:10 crc kubenswrapper[4646]: I1203 11:07:10.529294 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-vxhmd" Dec 03 11:07:10 crc kubenswrapper[4646]: I1203 11:07:10.635580 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-vxhmd" event={"ID":"72df546c-4ffc-46d9-9f79-54bb788a08fc","Type":"ContainerStarted","Data":"6f9ec339317ec6fadef89bf12adbca44be9f563f12c727ec5fc52cce72bcb5ff"} Dec 03 11:07:11 crc kubenswrapper[4646]: I1203 11:07:11.645320 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-vxhmd" event={"ID":"72df546c-4ffc-46d9-9f79-54bb788a08fc","Type":"ContainerStarted","Data":"16a73b9ae5f876f7ed631d90c8f941c8ea0f8fc936034f5b5cc962e0c9bbc5da"} Dec 03 11:07:14 crc kubenswrapper[4646]: I1203 11:07:14.663991 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-vxhmd" event={"ID":"72df546c-4ffc-46d9-9f79-54bb788a08fc","Type":"ContainerStarted","Data":"d7999ff31df55cd2b66e32de07a94e91c0bb4162369eee4128b33448c7fc0d6c"} Dec 03 11:07:15 crc kubenswrapper[4646]: I1203 11:07:15.669537 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-vxhmd" Dec 03 11:07:21 crc kubenswrapper[4646]: E1203 11:07:21.612108 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a" Dec 03 11:07:21 crc kubenswrapper[4646]: E1203 11:07:21.612825 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:frr-k8s-webhook-server,Image:registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a,Command:[/frr-k8s],Args:[--log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=$(NAMESPACE) --metrics-bind-address=:7572],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:monitoring,HostPort:0,ContainerPort:7572,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j4rgs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/metrics,Port:{1 0 monitoring},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/metrics,Port:{1 0 monitoring},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000700000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod frr-k8s-webhook-server-7fcb986d4-2ff9f_metallb-system(c5af6439-97a5-487d-b4a2-0fd0660af159): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 11:07:21 crc kubenswrapper[4646]: E1203 11:07:21.614003 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"frr-k8s-webhook-server\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" podUID="c5af6439-97a5-487d-b4a2-0fd0660af159" Dec 03 11:07:21 crc kubenswrapper[4646]: E1203 11:07:21.747110 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"frr-k8s-webhook-server\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a\\\"\"" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" podUID="c5af6439-97a5-487d-b4a2-0fd0660af159" Dec 03 11:07:21 crc kubenswrapper[4646]: I1203 11:07:21.764651 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-vxhmd" podStartSLOduration=15.764634079 podStartE2EDuration="15.764634079s" podCreationTimestamp="2025-12-03 11:07:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:07:15.707878902 +0000 UTC m=+812.170935037" watchObservedRunningTime="2025-12-03 11:07:21.764634079 +0000 UTC m=+818.227690214" Dec 03 11:07:24 crc kubenswrapper[4646]: I1203 11:07:24.766658 4646 generic.go:334] "Generic (PLEG): container finished" podID="a44776c2-244d-4ba3-b2f5-323eef5c8b9f" containerID="2293c8b470b992c7efda26215454f4483e13f33193145ca674e06a9a221b2358" exitCode=0 Dec 03 11:07:24 crc kubenswrapper[4646]: I1203 11:07:24.766753 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-knlr4" event={"ID":"a44776c2-244d-4ba3-b2f5-323eef5c8b9f","Type":"ContainerDied","Data":"2293c8b470b992c7efda26215454f4483e13f33193145ca674e06a9a221b2358"} Dec 03 11:07:25 crc kubenswrapper[4646]: I1203 11:07:25.775115 4646 generic.go:334] "Generic (PLEG): container finished" podID="a44776c2-244d-4ba3-b2f5-323eef5c8b9f" containerID="67e6851b62836244a3ecf81c9ed9029f8c33a8405958ac18c0d810d524eb024e" exitCode=0 Dec 03 11:07:25 crc kubenswrapper[4646]: I1203 11:07:25.775162 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-knlr4" event={"ID":"a44776c2-244d-4ba3-b2f5-323eef5c8b9f","Type":"ContainerDied","Data":"67e6851b62836244a3ecf81c9ed9029f8c33a8405958ac18c0d810d524eb024e"} Dec 03 11:07:25 crc kubenswrapper[4646]: I1203 11:07:25.964297 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:07:25 crc kubenswrapper[4646]: I1203 11:07:25.964377 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:07:26 crc kubenswrapper[4646]: I1203 11:07:26.784957 4646 generic.go:334] "Generic (PLEG): container finished" podID="a44776c2-244d-4ba3-b2f5-323eef5c8b9f" containerID="0cdcdc680cd00285bde6b56254b8f7af6c54b19b963715a2f539cbd4aafd667e" exitCode=0 Dec 03 11:07:26 crc kubenswrapper[4646]: I1203 11:07:26.785006 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-knlr4" event={"ID":"a44776c2-244d-4ba3-b2f5-323eef5c8b9f","Type":"ContainerDied","Data":"0cdcdc680cd00285bde6b56254b8f7af6c54b19b963715a2f539cbd4aafd667e"} Dec 03 11:07:26 crc kubenswrapper[4646]: I1203 11:07:26.962121 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-jfnxz" Dec 03 11:07:27 crc kubenswrapper[4646]: I1203 11:07:27.795894 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-knlr4" event={"ID":"a44776c2-244d-4ba3-b2f5-323eef5c8b9f","Type":"ContainerStarted","Data":"c33480ce36d2782d8aea6bc34d8c87e71bfa6b0a7963325bf126af7fab9edc5e"} Dec 03 11:07:27 crc kubenswrapper[4646]: I1203 11:07:27.795943 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-knlr4" event={"ID":"a44776c2-244d-4ba3-b2f5-323eef5c8b9f","Type":"ContainerStarted","Data":"f7301e3d8127358459259ed8219a507e17f6452b01ed5acb480d23a5ac5fb403"} Dec 03 11:07:27 crc kubenswrapper[4646]: I1203 11:07:27.795967 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-knlr4" event={"ID":"a44776c2-244d-4ba3-b2f5-323eef5c8b9f","Type":"ContainerStarted","Data":"5e8f16d9fb79d4813f846a44e604f7dd3dddb5a3ed2c651423dda749a43be83f"} Dec 03 11:07:28 crc kubenswrapper[4646]: I1203 11:07:28.806310 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-knlr4" event={"ID":"a44776c2-244d-4ba3-b2f5-323eef5c8b9f","Type":"ContainerStarted","Data":"91b967ba5849ba17bf09dd2ddbec7b4642debc406fbf07125da6ef82ee66e9fa"} Dec 03 11:07:28 crc kubenswrapper[4646]: I1203 11:07:28.806683 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-knlr4" event={"ID":"a44776c2-244d-4ba3-b2f5-323eef5c8b9f","Type":"ContainerStarted","Data":"271674bb951543aa69d9509054b45d2aa45a894d97b808ff7db4720f82fd46b3"} Dec 03 11:07:28 crc kubenswrapper[4646]: I1203 11:07:28.806699 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-knlr4" event={"ID":"a44776c2-244d-4ba3-b2f5-323eef5c8b9f","Type":"ContainerStarted","Data":"28afb0a4b7834d214213b845b685a0a31b3af76e057a66fdfb7af8fa3840f50c"} Dec 03 11:07:28 crc kubenswrapper[4646]: I1203 11:07:28.806818 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:28 crc kubenswrapper[4646]: I1203 11:07:28.835394 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-knlr4" podStartSLOduration=7.515379751 podStartE2EDuration="22.835293996s" podCreationTimestamp="2025-12-03 11:07:06 +0000 UTC" firstStartedPulling="2025-12-03 11:07:08.471192927 +0000 UTC m=+804.934249072" lastFinishedPulling="2025-12-03 11:07:23.791107182 +0000 UTC m=+820.254163317" observedRunningTime="2025-12-03 11:07:28.830262956 +0000 UTC m=+825.293319091" watchObservedRunningTime="2025-12-03 11:07:28.835293996 +0000 UTC m=+825.298350131" Dec 03 11:07:30 crc kubenswrapper[4646]: I1203 11:07:30.532813 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-vxhmd" Dec 03 11:07:32 crc kubenswrapper[4646]: I1203 11:07:32.424129 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:32 crc kubenswrapper[4646]: I1203 11:07:32.459513 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:33 crc kubenswrapper[4646]: I1203 11:07:33.197065 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-7wk9g"] Dec 03 11:07:33 crc kubenswrapper[4646]: I1203 11:07:33.198675 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7wk9g" Dec 03 11:07:33 crc kubenswrapper[4646]: I1203 11:07:33.206049 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-47t9v" Dec 03 11:07:33 crc kubenswrapper[4646]: I1203 11:07:33.206367 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 03 11:07:33 crc kubenswrapper[4646]: I1203 11:07:33.206412 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 03 11:07:33 crc kubenswrapper[4646]: I1203 11:07:33.221216 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7wk9g"] Dec 03 11:07:33 crc kubenswrapper[4646]: I1203 11:07:33.361411 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg6ff\" (UniqueName: \"kubernetes.io/projected/f4a317f7-f6a9-4b6e-9a32-28477211d8f1-kube-api-access-xg6ff\") pod \"openstack-operator-index-7wk9g\" (UID: \"f4a317f7-f6a9-4b6e-9a32-28477211d8f1\") " pod="openstack-operators/openstack-operator-index-7wk9g" Dec 03 11:07:33 crc kubenswrapper[4646]: I1203 11:07:33.462634 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg6ff\" (UniqueName: \"kubernetes.io/projected/f4a317f7-f6a9-4b6e-9a32-28477211d8f1-kube-api-access-xg6ff\") pod \"openstack-operator-index-7wk9g\" (UID: \"f4a317f7-f6a9-4b6e-9a32-28477211d8f1\") " pod="openstack-operators/openstack-operator-index-7wk9g" Dec 03 11:07:33 crc kubenswrapper[4646]: I1203 11:07:33.487452 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg6ff\" (UniqueName: \"kubernetes.io/projected/f4a317f7-f6a9-4b6e-9a32-28477211d8f1-kube-api-access-xg6ff\") pod \"openstack-operator-index-7wk9g\" (UID: \"f4a317f7-f6a9-4b6e-9a32-28477211d8f1\") " pod="openstack-operators/openstack-operator-index-7wk9g" Dec 03 11:07:33 crc kubenswrapper[4646]: I1203 11:07:33.517918 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7wk9g" Dec 03 11:07:34 crc kubenswrapper[4646]: I1203 11:07:34.141230 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7wk9g"] Dec 03 11:07:34 crc kubenswrapper[4646]: I1203 11:07:34.842222 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7wk9g" event={"ID":"f4a317f7-f6a9-4b6e-9a32-28477211d8f1","Type":"ContainerStarted","Data":"a666198e410b44e54b811e0e44cc039348b197a10095724f626b631e35382de0"} Dec 03 11:07:35 crc kubenswrapper[4646]: I1203 11:07:35.848113 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" event={"ID":"c5af6439-97a5-487d-b4a2-0fd0660af159","Type":"ContainerStarted","Data":"3aa8b07d99cbcbe56bd4a0498929b3f188f1861111796976d768742d181430a9"} Dec 03 11:07:35 crc kubenswrapper[4646]: I1203 11:07:35.849608 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" Dec 03 11:07:35 crc kubenswrapper[4646]: I1203 11:07:35.868846 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" podStartSLOduration=-9223372006.985947 podStartE2EDuration="29.868828797s" podCreationTimestamp="2025-12-03 11:07:06 +0000 UTC" firstStartedPulling="2025-12-03 11:07:07.651801188 +0000 UTC m=+804.114857323" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:07:35.863186939 +0000 UTC m=+832.326243074" watchObservedRunningTime="2025-12-03 11:07:35.868828797 +0000 UTC m=+832.331884932" Dec 03 11:07:37 crc kubenswrapper[4646]: I1203 11:07:37.426664 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-knlr4" Dec 03 11:07:37 crc kubenswrapper[4646]: I1203 11:07:37.565608 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7wk9g"] Dec 03 11:07:37 crc kubenswrapper[4646]: I1203 11:07:37.973471 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zdrkq"] Dec 03 11:07:37 crc kubenswrapper[4646]: I1203 11:07:37.974242 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zdrkq" Dec 03 11:07:37 crc kubenswrapper[4646]: I1203 11:07:37.993656 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zdrkq"] Dec 03 11:07:38 crc kubenswrapper[4646]: I1203 11:07:38.071971 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9v9m\" (UniqueName: \"kubernetes.io/projected/a15c5248-9b54-4744-b8e9-f411b6a3c1a1-kube-api-access-z9v9m\") pod \"openstack-operator-index-zdrkq\" (UID: \"a15c5248-9b54-4744-b8e9-f411b6a3c1a1\") " pod="openstack-operators/openstack-operator-index-zdrkq" Dec 03 11:07:38 crc kubenswrapper[4646]: I1203 11:07:38.172870 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9v9m\" (UniqueName: \"kubernetes.io/projected/a15c5248-9b54-4744-b8e9-f411b6a3c1a1-kube-api-access-z9v9m\") pod \"openstack-operator-index-zdrkq\" (UID: \"a15c5248-9b54-4744-b8e9-f411b6a3c1a1\") " pod="openstack-operators/openstack-operator-index-zdrkq" Dec 03 11:07:38 crc kubenswrapper[4646]: I1203 11:07:38.190913 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9v9m\" (UniqueName: \"kubernetes.io/projected/a15c5248-9b54-4744-b8e9-f411b6a3c1a1-kube-api-access-z9v9m\") pod \"openstack-operator-index-zdrkq\" (UID: \"a15c5248-9b54-4744-b8e9-f411b6a3c1a1\") " pod="openstack-operators/openstack-operator-index-zdrkq" Dec 03 11:07:38 crc kubenswrapper[4646]: I1203 11:07:38.297920 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zdrkq" Dec 03 11:07:45 crc kubenswrapper[4646]: I1203 11:07:45.558057 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zdrkq"] Dec 03 11:07:45 crc kubenswrapper[4646]: I1203 11:07:45.908833 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zdrkq" event={"ID":"a15c5248-9b54-4744-b8e9-f411b6a3c1a1","Type":"ContainerStarted","Data":"a8fb2938b301e168354711580434c45acf6b8cd0f511e0db37e64e88fca9c689"} Dec 03 11:07:47 crc kubenswrapper[4646]: I1203 11:07:47.441531 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-2ff9f" Dec 03 11:07:48 crc kubenswrapper[4646]: I1203 11:07:48.934645 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7wk9g" event={"ID":"f4a317f7-f6a9-4b6e-9a32-28477211d8f1","Type":"ContainerStarted","Data":"fc90f673f7f50e007eba8ab21faa30939c205da775f964dc6032c29d8764f9f8"} Dec 03 11:07:49 crc kubenswrapper[4646]: I1203 11:07:49.942262 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zdrkq" event={"ID":"a15c5248-9b54-4744-b8e9-f411b6a3c1a1","Type":"ContainerStarted","Data":"acd9aad06e708073254a6a61199f40051e80abc56c218c4f89bcf61b4bd924b6"} Dec 03 11:07:49 crc kubenswrapper[4646]: I1203 11:07:49.942891 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-7wk9g" podUID="f4a317f7-f6a9-4b6e-9a32-28477211d8f1" containerName="registry-server" containerID="cri-o://fc90f673f7f50e007eba8ab21faa30939c205da775f964dc6032c29d8764f9f8" gracePeriod=2 Dec 03 11:07:49 crc kubenswrapper[4646]: I1203 11:07:49.965500 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zdrkq" podStartSLOduration=10.05415597 podStartE2EDuration="12.965437308s" podCreationTimestamp="2025-12-03 11:07:37 +0000 UTC" firstStartedPulling="2025-12-03 11:07:45.679117611 +0000 UTC m=+842.142173746" lastFinishedPulling="2025-12-03 11:07:48.590398949 +0000 UTC m=+845.053455084" observedRunningTime="2025-12-03 11:07:49.962233066 +0000 UTC m=+846.425289211" watchObservedRunningTime="2025-12-03 11:07:49.965437308 +0000 UTC m=+846.428493473" Dec 03 11:07:49 crc kubenswrapper[4646]: I1203 11:07:49.984888 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-7wk9g" podStartSLOduration=2.609264951 podStartE2EDuration="16.984867828s" podCreationTimestamp="2025-12-03 11:07:33 +0000 UTC" firstStartedPulling="2025-12-03 11:07:34.161642881 +0000 UTC m=+830.624699016" lastFinishedPulling="2025-12-03 11:07:48.537245758 +0000 UTC m=+845.000301893" observedRunningTime="2025-12-03 11:07:49.982329294 +0000 UTC m=+846.445385419" watchObservedRunningTime="2025-12-03 11:07:49.984867828 +0000 UTC m=+846.447923963" Dec 03 11:07:50 crc kubenswrapper[4646]: I1203 11:07:50.369214 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7wk9g" Dec 03 11:07:50 crc kubenswrapper[4646]: I1203 11:07:50.442852 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xg6ff\" (UniqueName: \"kubernetes.io/projected/f4a317f7-f6a9-4b6e-9a32-28477211d8f1-kube-api-access-xg6ff\") pod \"f4a317f7-f6a9-4b6e-9a32-28477211d8f1\" (UID: \"f4a317f7-f6a9-4b6e-9a32-28477211d8f1\") " Dec 03 11:07:50 crc kubenswrapper[4646]: I1203 11:07:50.448068 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4a317f7-f6a9-4b6e-9a32-28477211d8f1-kube-api-access-xg6ff" (OuterVolumeSpecName: "kube-api-access-xg6ff") pod "f4a317f7-f6a9-4b6e-9a32-28477211d8f1" (UID: "f4a317f7-f6a9-4b6e-9a32-28477211d8f1"). InnerVolumeSpecName "kube-api-access-xg6ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:07:50 crc kubenswrapper[4646]: I1203 11:07:50.544256 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xg6ff\" (UniqueName: \"kubernetes.io/projected/f4a317f7-f6a9-4b6e-9a32-28477211d8f1-kube-api-access-xg6ff\") on node \"crc\" DevicePath \"\"" Dec 03 11:07:50 crc kubenswrapper[4646]: I1203 11:07:50.964614 4646 generic.go:334] "Generic (PLEG): container finished" podID="f4a317f7-f6a9-4b6e-9a32-28477211d8f1" containerID="fc90f673f7f50e007eba8ab21faa30939c205da775f964dc6032c29d8764f9f8" exitCode=0 Dec 03 11:07:50 crc kubenswrapper[4646]: I1203 11:07:50.964728 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7wk9g" Dec 03 11:07:50 crc kubenswrapper[4646]: I1203 11:07:50.964746 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7wk9g" event={"ID":"f4a317f7-f6a9-4b6e-9a32-28477211d8f1","Type":"ContainerDied","Data":"fc90f673f7f50e007eba8ab21faa30939c205da775f964dc6032c29d8764f9f8"} Dec 03 11:07:50 crc kubenswrapper[4646]: I1203 11:07:50.964906 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7wk9g" event={"ID":"f4a317f7-f6a9-4b6e-9a32-28477211d8f1","Type":"ContainerDied","Data":"a666198e410b44e54b811e0e44cc039348b197a10095724f626b631e35382de0"} Dec 03 11:07:50 crc kubenswrapper[4646]: I1203 11:07:50.964948 4646 scope.go:117] "RemoveContainer" containerID="fc90f673f7f50e007eba8ab21faa30939c205da775f964dc6032c29d8764f9f8" Dec 03 11:07:50 crc kubenswrapper[4646]: I1203 11:07:50.987976 4646 scope.go:117] "RemoveContainer" containerID="fc90f673f7f50e007eba8ab21faa30939c205da775f964dc6032c29d8764f9f8" Dec 03 11:07:50 crc kubenswrapper[4646]: E1203 11:07:50.988747 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc90f673f7f50e007eba8ab21faa30939c205da775f964dc6032c29d8764f9f8\": container with ID starting with fc90f673f7f50e007eba8ab21faa30939c205da775f964dc6032c29d8764f9f8 not found: ID does not exist" containerID="fc90f673f7f50e007eba8ab21faa30939c205da775f964dc6032c29d8764f9f8" Dec 03 11:07:50 crc kubenswrapper[4646]: I1203 11:07:50.988798 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc90f673f7f50e007eba8ab21faa30939c205da775f964dc6032c29d8764f9f8"} err="failed to get container status \"fc90f673f7f50e007eba8ab21faa30939c205da775f964dc6032c29d8764f9f8\": rpc error: code = NotFound desc = could not find container \"fc90f673f7f50e007eba8ab21faa30939c205da775f964dc6032c29d8764f9f8\": container with ID starting with fc90f673f7f50e007eba8ab21faa30939c205da775f964dc6032c29d8764f9f8 not found: ID does not exist" Dec 03 11:07:51 crc kubenswrapper[4646]: I1203 11:07:51.002100 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7wk9g"] Dec 03 11:07:51 crc kubenswrapper[4646]: I1203 11:07:51.006077 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-7wk9g"] Dec 03 11:07:51 crc kubenswrapper[4646]: I1203 11:07:51.856136 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4a317f7-f6a9-4b6e-9a32-28477211d8f1" path="/var/lib/kubelet/pods/f4a317f7-f6a9-4b6e-9a32-28477211d8f1/volumes" Dec 03 11:07:55 crc kubenswrapper[4646]: I1203 11:07:55.964067 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:07:55 crc kubenswrapper[4646]: I1203 11:07:55.964620 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:07:55 crc kubenswrapper[4646]: I1203 11:07:55.964683 4646 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 11:07:55 crc kubenswrapper[4646]: I1203 11:07:55.965285 4646 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"41b509e57c2eee648a101eefda70c1101098be6f2f4ca7c3a458c795b0cc5f55"} pod="openshift-machine-config-operator/machine-config-daemon-pndzt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 11:07:55 crc kubenswrapper[4646]: I1203 11:07:55.965366 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" containerID="cri-o://41b509e57c2eee648a101eefda70c1101098be6f2f4ca7c3a458c795b0cc5f55" gracePeriod=600 Dec 03 11:07:57 crc kubenswrapper[4646]: I1203 11:07:57.004109 4646 generic.go:334] "Generic (PLEG): container finished" podID="9695d947-fc43-4625-86af-a1a3e177bb27" containerID="41b509e57c2eee648a101eefda70c1101098be6f2f4ca7c3a458c795b0cc5f55" exitCode=0 Dec 03 11:07:57 crc kubenswrapper[4646]: I1203 11:07:57.004161 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerDied","Data":"41b509e57c2eee648a101eefda70c1101098be6f2f4ca7c3a458c795b0cc5f55"} Dec 03 11:07:57 crc kubenswrapper[4646]: I1203 11:07:57.004432 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerStarted","Data":"14e17deb53ff7184afa03eb1fff9074fe416c1136352cde34645028a19901f16"} Dec 03 11:07:57 crc kubenswrapper[4646]: I1203 11:07:57.004454 4646 scope.go:117] "RemoveContainer" containerID="bc5af8932094babb5f9581e57117493e897a2b27daa913a38481a13fa5135b5e" Dec 03 11:07:58 crc kubenswrapper[4646]: I1203 11:07:58.298297 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-zdrkq" Dec 03 11:07:58 crc kubenswrapper[4646]: I1203 11:07:58.298701 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-zdrkq" Dec 03 11:07:58 crc kubenswrapper[4646]: I1203 11:07:58.324854 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-zdrkq" Dec 03 11:07:59 crc kubenswrapper[4646]: I1203 11:07:59.054448 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-zdrkq" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.312157 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs"] Dec 03 11:08:05 crc kubenswrapper[4646]: E1203 11:08:05.312888 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a317f7-f6a9-4b6e-9a32-28477211d8f1" containerName="registry-server" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.312899 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a317f7-f6a9-4b6e-9a32-28477211d8f1" containerName="registry-server" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.313007 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4a317f7-f6a9-4b6e-9a32-28477211d8f1" containerName="registry-server" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.313811 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.317392 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-4t8wr" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.325747 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs"] Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.326851 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5hpr\" (UniqueName: \"kubernetes.io/projected/f9512558-d306-43f1-b530-6c25875a7625-kube-api-access-q5hpr\") pod \"93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs\" (UID: \"f9512558-d306-43f1-b530-6c25875a7625\") " pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.326924 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f9512558-d306-43f1-b530-6c25875a7625-bundle\") pod \"93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs\" (UID: \"f9512558-d306-43f1-b530-6c25875a7625\") " pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.326951 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f9512558-d306-43f1-b530-6c25875a7625-util\") pod \"93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs\" (UID: \"f9512558-d306-43f1-b530-6c25875a7625\") " pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.428240 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5hpr\" (UniqueName: \"kubernetes.io/projected/f9512558-d306-43f1-b530-6c25875a7625-kube-api-access-q5hpr\") pod \"93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs\" (UID: \"f9512558-d306-43f1-b530-6c25875a7625\") " pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.428530 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f9512558-d306-43f1-b530-6c25875a7625-bundle\") pod \"93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs\" (UID: \"f9512558-d306-43f1-b530-6c25875a7625\") " pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.428656 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f9512558-d306-43f1-b530-6c25875a7625-util\") pod \"93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs\" (UID: \"f9512558-d306-43f1-b530-6c25875a7625\") " pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.429046 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f9512558-d306-43f1-b530-6c25875a7625-bundle\") pod \"93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs\" (UID: \"f9512558-d306-43f1-b530-6c25875a7625\") " pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.429089 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f9512558-d306-43f1-b530-6c25875a7625-util\") pod \"93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs\" (UID: \"f9512558-d306-43f1-b530-6c25875a7625\") " pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.446666 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5hpr\" (UniqueName: \"kubernetes.io/projected/f9512558-d306-43f1-b530-6c25875a7625-kube-api-access-q5hpr\") pod \"93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs\" (UID: \"f9512558-d306-43f1-b530-6c25875a7625\") " pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.631067 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" Dec 03 11:08:05 crc kubenswrapper[4646]: I1203 11:08:05.888380 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs"] Dec 03 11:08:06 crc kubenswrapper[4646]: I1203 11:08:06.062195 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" event={"ID":"f9512558-d306-43f1-b530-6c25875a7625","Type":"ContainerStarted","Data":"9b037f89faae7542910cc0f79d6484b58482a697586ca123f36c89f46e338cf5"} Dec 03 11:08:07 crc kubenswrapper[4646]: I1203 11:08:07.070240 4646 generic.go:334] "Generic (PLEG): container finished" podID="f9512558-d306-43f1-b530-6c25875a7625" containerID="7e95eaf7f26fcfb8bb15cb79ca53c9336ed333161b6d774e92905e75e886ce18" exitCode=0 Dec 03 11:08:07 crc kubenswrapper[4646]: I1203 11:08:07.070355 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" event={"ID":"f9512558-d306-43f1-b530-6c25875a7625","Type":"ContainerDied","Data":"7e95eaf7f26fcfb8bb15cb79ca53c9336ed333161b6d774e92905e75e886ce18"} Dec 03 11:08:08 crc kubenswrapper[4646]: I1203 11:08:08.076872 4646 generic.go:334] "Generic (PLEG): container finished" podID="f9512558-d306-43f1-b530-6c25875a7625" containerID="40393670507c6e2ab6b934119b428d50a76cfe24f209fd7855fcb1422f72ee83" exitCode=0 Dec 03 11:08:08 crc kubenswrapper[4646]: I1203 11:08:08.076967 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" event={"ID":"f9512558-d306-43f1-b530-6c25875a7625","Type":"ContainerDied","Data":"40393670507c6e2ab6b934119b428d50a76cfe24f209fd7855fcb1422f72ee83"} Dec 03 11:08:09 crc kubenswrapper[4646]: I1203 11:08:09.086719 4646 generic.go:334] "Generic (PLEG): container finished" podID="f9512558-d306-43f1-b530-6c25875a7625" containerID="980ad58aa92cd0261c3546bff39b1727749932bdcdf249207ee4ce4c77eeab7b" exitCode=0 Dec 03 11:08:09 crc kubenswrapper[4646]: I1203 11:08:09.086796 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" event={"ID":"f9512558-d306-43f1-b530-6c25875a7625","Type":"ContainerDied","Data":"980ad58aa92cd0261c3546bff39b1727749932bdcdf249207ee4ce4c77eeab7b"} Dec 03 11:08:10 crc kubenswrapper[4646]: I1203 11:08:10.366136 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" Dec 03 11:08:10 crc kubenswrapper[4646]: I1203 11:08:10.494903 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f9512558-d306-43f1-b530-6c25875a7625-bundle\") pod \"f9512558-d306-43f1-b530-6c25875a7625\" (UID: \"f9512558-d306-43f1-b530-6c25875a7625\") " Dec 03 11:08:10 crc kubenswrapper[4646]: I1203 11:08:10.494996 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f9512558-d306-43f1-b530-6c25875a7625-util\") pod \"f9512558-d306-43f1-b530-6c25875a7625\" (UID: \"f9512558-d306-43f1-b530-6c25875a7625\") " Dec 03 11:08:10 crc kubenswrapper[4646]: I1203 11:08:10.495160 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5hpr\" (UniqueName: \"kubernetes.io/projected/f9512558-d306-43f1-b530-6c25875a7625-kube-api-access-q5hpr\") pod \"f9512558-d306-43f1-b530-6c25875a7625\" (UID: \"f9512558-d306-43f1-b530-6c25875a7625\") " Dec 03 11:08:10 crc kubenswrapper[4646]: I1203 11:08:10.496488 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9512558-d306-43f1-b530-6c25875a7625-bundle" (OuterVolumeSpecName: "bundle") pod "f9512558-d306-43f1-b530-6c25875a7625" (UID: "f9512558-d306-43f1-b530-6c25875a7625"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:08:10 crc kubenswrapper[4646]: I1203 11:08:10.501251 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9512558-d306-43f1-b530-6c25875a7625-kube-api-access-q5hpr" (OuterVolumeSpecName: "kube-api-access-q5hpr") pod "f9512558-d306-43f1-b530-6c25875a7625" (UID: "f9512558-d306-43f1-b530-6c25875a7625"). InnerVolumeSpecName "kube-api-access-q5hpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:08:10 crc kubenswrapper[4646]: I1203 11:08:10.509994 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9512558-d306-43f1-b530-6c25875a7625-util" (OuterVolumeSpecName: "util") pod "f9512558-d306-43f1-b530-6c25875a7625" (UID: "f9512558-d306-43f1-b530-6c25875a7625"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:08:10 crc kubenswrapper[4646]: I1203 11:08:10.597167 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5hpr\" (UniqueName: \"kubernetes.io/projected/f9512558-d306-43f1-b530-6c25875a7625-kube-api-access-q5hpr\") on node \"crc\" DevicePath \"\"" Dec 03 11:08:10 crc kubenswrapper[4646]: I1203 11:08:10.597233 4646 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f9512558-d306-43f1-b530-6c25875a7625-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:08:10 crc kubenswrapper[4646]: I1203 11:08:10.597242 4646 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f9512558-d306-43f1-b530-6c25875a7625-util\") on node \"crc\" DevicePath \"\"" Dec 03 11:08:11 crc kubenswrapper[4646]: I1203 11:08:11.106830 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" event={"ID":"f9512558-d306-43f1-b530-6c25875a7625","Type":"ContainerDied","Data":"9b037f89faae7542910cc0f79d6484b58482a697586ca123f36c89f46e338cf5"} Dec 03 11:08:11 crc kubenswrapper[4646]: I1203 11:08:11.106876 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b037f89faae7542910cc0f79d6484b58482a697586ca123f36c89f46e338cf5" Dec 03 11:08:11 crc kubenswrapper[4646]: I1203 11:08:11.106899 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs" Dec 03 11:08:17 crc kubenswrapper[4646]: I1203 11:08:17.291553 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-779dc79ddf-9gqzp"] Dec 03 11:08:17 crc kubenswrapper[4646]: E1203 11:08:17.292407 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9512558-d306-43f1-b530-6c25875a7625" containerName="util" Dec 03 11:08:17 crc kubenswrapper[4646]: I1203 11:08:17.292423 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9512558-d306-43f1-b530-6c25875a7625" containerName="util" Dec 03 11:08:17 crc kubenswrapper[4646]: E1203 11:08:17.292439 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9512558-d306-43f1-b530-6c25875a7625" containerName="extract" Dec 03 11:08:17 crc kubenswrapper[4646]: I1203 11:08:17.292446 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9512558-d306-43f1-b530-6c25875a7625" containerName="extract" Dec 03 11:08:17 crc kubenswrapper[4646]: E1203 11:08:17.292460 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9512558-d306-43f1-b530-6c25875a7625" containerName="pull" Dec 03 11:08:17 crc kubenswrapper[4646]: I1203 11:08:17.292468 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9512558-d306-43f1-b530-6c25875a7625" containerName="pull" Dec 03 11:08:17 crc kubenswrapper[4646]: I1203 11:08:17.292598 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9512558-d306-43f1-b530-6c25875a7625" containerName="extract" Dec 03 11:08:17 crc kubenswrapper[4646]: I1203 11:08:17.293065 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-779dc79ddf-9gqzp" Dec 03 11:08:17 crc kubenswrapper[4646]: W1203 11:08:17.296315 4646 reflector.go:561] object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-rs6sv": failed to list *v1.Secret: secrets "openstack-operator-controller-operator-dockercfg-rs6sv" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack-operators": no relationship found between node 'crc' and this object Dec 03 11:08:17 crc kubenswrapper[4646]: E1203 11:08:17.296380 4646 reflector.go:158] "Unhandled Error" err="object-\"openstack-operators\"/\"openstack-operator-controller-operator-dockercfg-rs6sv\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openstack-operator-controller-operator-dockercfg-rs6sv\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 11:08:17 crc kubenswrapper[4646]: I1203 11:08:17.297769 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdwbm\" (UniqueName: \"kubernetes.io/projected/509a9181-c788-464c-b157-4928132fadbe-kube-api-access-hdwbm\") pod \"openstack-operator-controller-operator-779dc79ddf-9gqzp\" (UID: \"509a9181-c788-464c-b157-4928132fadbe\") " pod="openstack-operators/openstack-operator-controller-operator-779dc79ddf-9gqzp" Dec 03 11:08:17 crc kubenswrapper[4646]: I1203 11:08:17.326050 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-779dc79ddf-9gqzp"] Dec 03 11:08:17 crc kubenswrapper[4646]: I1203 11:08:17.399373 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdwbm\" (UniqueName: \"kubernetes.io/projected/509a9181-c788-464c-b157-4928132fadbe-kube-api-access-hdwbm\") pod \"openstack-operator-controller-operator-779dc79ddf-9gqzp\" (UID: \"509a9181-c788-464c-b157-4928132fadbe\") " pod="openstack-operators/openstack-operator-controller-operator-779dc79ddf-9gqzp" Dec 03 11:08:17 crc kubenswrapper[4646]: I1203 11:08:17.428726 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdwbm\" (UniqueName: \"kubernetes.io/projected/509a9181-c788-464c-b157-4928132fadbe-kube-api-access-hdwbm\") pod \"openstack-operator-controller-operator-779dc79ddf-9gqzp\" (UID: \"509a9181-c788-464c-b157-4928132fadbe\") " pod="openstack-operators/openstack-operator-controller-operator-779dc79ddf-9gqzp" Dec 03 11:08:18 crc kubenswrapper[4646]: I1203 11:08:18.408130 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-rs6sv" Dec 03 11:08:18 crc kubenswrapper[4646]: I1203 11:08:18.410655 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-779dc79ddf-9gqzp" Dec 03 11:08:18 crc kubenswrapper[4646]: I1203 11:08:18.624048 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-779dc79ddf-9gqzp"] Dec 03 11:08:19 crc kubenswrapper[4646]: I1203 11:08:19.153328 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-779dc79ddf-9gqzp" event={"ID":"509a9181-c788-464c-b157-4928132fadbe","Type":"ContainerStarted","Data":"9a65db04da8aaaba9589676f6a42c130e55841dbe4524471dd858c91db59083c"} Dec 03 11:08:26 crc kubenswrapper[4646]: I1203 11:08:26.775165 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-779dc79ddf-9gqzp" event={"ID":"509a9181-c788-464c-b157-4928132fadbe","Type":"ContainerStarted","Data":"d103aefa1f196d50d701483d1d9ba83c65586efcaf8678b8ad5a6a1f3823197b"} Dec 03 11:08:26 crc kubenswrapper[4646]: I1203 11:08:26.776422 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-779dc79ddf-9gqzp" Dec 03 11:08:26 crc kubenswrapper[4646]: I1203 11:08:26.807801 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-779dc79ddf-9gqzp" podStartSLOduration=2.024410212 podStartE2EDuration="9.807783171s" podCreationTimestamp="2025-12-03 11:08:17 +0000 UTC" firstStartedPulling="2025-12-03 11:08:18.635074777 +0000 UTC m=+875.098130912" lastFinishedPulling="2025-12-03 11:08:26.418447726 +0000 UTC m=+882.881503871" observedRunningTime="2025-12-03 11:08:26.803941591 +0000 UTC m=+883.266997726" watchObservedRunningTime="2025-12-03 11:08:26.807783171 +0000 UTC m=+883.270839306" Dec 03 11:08:38 crc kubenswrapper[4646]: I1203 11:08:38.412850 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-779dc79ddf-9gqzp" Dec 03 11:08:45 crc kubenswrapper[4646]: I1203 11:08:45.144625 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qspkq"] Dec 03 11:08:45 crc kubenswrapper[4646]: I1203 11:08:45.146731 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qspkq" Dec 03 11:08:45 crc kubenswrapper[4646]: I1203 11:08:45.201274 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qspkq"] Dec 03 11:08:45 crc kubenswrapper[4646]: I1203 11:08:45.273690 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/682649bd-4003-4e88-a81a-dafee3703114-catalog-content\") pod \"certified-operators-qspkq\" (UID: \"682649bd-4003-4e88-a81a-dafee3703114\") " pod="openshift-marketplace/certified-operators-qspkq" Dec 03 11:08:45 crc kubenswrapper[4646]: I1203 11:08:45.273775 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/682649bd-4003-4e88-a81a-dafee3703114-utilities\") pod \"certified-operators-qspkq\" (UID: \"682649bd-4003-4e88-a81a-dafee3703114\") " pod="openshift-marketplace/certified-operators-qspkq" Dec 03 11:08:45 crc kubenswrapper[4646]: I1203 11:08:45.273871 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46rm9\" (UniqueName: \"kubernetes.io/projected/682649bd-4003-4e88-a81a-dafee3703114-kube-api-access-46rm9\") pod \"certified-operators-qspkq\" (UID: \"682649bd-4003-4e88-a81a-dafee3703114\") " pod="openshift-marketplace/certified-operators-qspkq" Dec 03 11:08:45 crc kubenswrapper[4646]: I1203 11:08:45.375282 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/682649bd-4003-4e88-a81a-dafee3703114-utilities\") pod \"certified-operators-qspkq\" (UID: \"682649bd-4003-4e88-a81a-dafee3703114\") " pod="openshift-marketplace/certified-operators-qspkq" Dec 03 11:08:45 crc kubenswrapper[4646]: I1203 11:08:45.375396 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46rm9\" (UniqueName: \"kubernetes.io/projected/682649bd-4003-4e88-a81a-dafee3703114-kube-api-access-46rm9\") pod \"certified-operators-qspkq\" (UID: \"682649bd-4003-4e88-a81a-dafee3703114\") " pod="openshift-marketplace/certified-operators-qspkq" Dec 03 11:08:45 crc kubenswrapper[4646]: I1203 11:08:45.375421 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/682649bd-4003-4e88-a81a-dafee3703114-catalog-content\") pod \"certified-operators-qspkq\" (UID: \"682649bd-4003-4e88-a81a-dafee3703114\") " pod="openshift-marketplace/certified-operators-qspkq" Dec 03 11:08:45 crc kubenswrapper[4646]: I1203 11:08:45.375872 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/682649bd-4003-4e88-a81a-dafee3703114-utilities\") pod \"certified-operators-qspkq\" (UID: \"682649bd-4003-4e88-a81a-dafee3703114\") " pod="openshift-marketplace/certified-operators-qspkq" Dec 03 11:08:45 crc kubenswrapper[4646]: I1203 11:08:45.375925 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/682649bd-4003-4e88-a81a-dafee3703114-catalog-content\") pod \"certified-operators-qspkq\" (UID: \"682649bd-4003-4e88-a81a-dafee3703114\") " pod="openshift-marketplace/certified-operators-qspkq" Dec 03 11:08:45 crc kubenswrapper[4646]: I1203 11:08:45.418245 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46rm9\" (UniqueName: \"kubernetes.io/projected/682649bd-4003-4e88-a81a-dafee3703114-kube-api-access-46rm9\") pod \"certified-operators-qspkq\" (UID: \"682649bd-4003-4e88-a81a-dafee3703114\") " pod="openshift-marketplace/certified-operators-qspkq" Dec 03 11:08:45 crc kubenswrapper[4646]: I1203 11:08:45.465540 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qspkq" Dec 03 11:08:46 crc kubenswrapper[4646]: I1203 11:08:46.259284 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qspkq"] Dec 03 11:08:46 crc kubenswrapper[4646]: W1203 11:08:46.277627 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod682649bd_4003_4e88_a81a_dafee3703114.slice/crio-e4e9d4dca3c62d805e5d5f599a2ed5f82ad1d9705804aff328669d4020ed0154 WatchSource:0}: Error finding container e4e9d4dca3c62d805e5d5f599a2ed5f82ad1d9705804aff328669d4020ed0154: Status 404 returned error can't find the container with id e4e9d4dca3c62d805e5d5f599a2ed5f82ad1d9705804aff328669d4020ed0154 Dec 03 11:08:46 crc kubenswrapper[4646]: I1203 11:08:46.960455 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qspkq" event={"ID":"682649bd-4003-4e88-a81a-dafee3703114","Type":"ContainerStarted","Data":"e4e9d4dca3c62d805e5d5f599a2ed5f82ad1d9705804aff328669d4020ed0154"} Dec 03 11:08:47 crc kubenswrapper[4646]: I1203 11:08:47.967012 4646 generic.go:334] "Generic (PLEG): container finished" podID="682649bd-4003-4e88-a81a-dafee3703114" containerID="09688d1f00ee1d160c7ce6560dd46eaa961344874cf1b92bcee1d919dfd5bd1e" exitCode=0 Dec 03 11:08:47 crc kubenswrapper[4646]: I1203 11:08:47.967280 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qspkq" event={"ID":"682649bd-4003-4e88-a81a-dafee3703114","Type":"ContainerDied","Data":"09688d1f00ee1d160c7ce6560dd46eaa961344874cf1b92bcee1d919dfd5bd1e"} Dec 03 11:08:51 crc kubenswrapper[4646]: I1203 11:08:51.027241 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n8pb7"] Dec 03 11:08:51 crc kubenswrapper[4646]: I1203 11:08:51.028436 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:08:51 crc kubenswrapper[4646]: I1203 11:08:51.034879 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n8pb7"] Dec 03 11:08:51 crc kubenswrapper[4646]: I1203 11:08:51.036168 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjz45\" (UniqueName: \"kubernetes.io/projected/b414a389-9e48-412c-b4e0-26a9b4cb51eb-kube-api-access-hjz45\") pod \"community-operators-n8pb7\" (UID: \"b414a389-9e48-412c-b4e0-26a9b4cb51eb\") " pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:08:51 crc kubenswrapper[4646]: I1203 11:08:51.036231 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b414a389-9e48-412c-b4e0-26a9b4cb51eb-catalog-content\") pod \"community-operators-n8pb7\" (UID: \"b414a389-9e48-412c-b4e0-26a9b4cb51eb\") " pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:08:51 crc kubenswrapper[4646]: I1203 11:08:51.036266 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b414a389-9e48-412c-b4e0-26a9b4cb51eb-utilities\") pod \"community-operators-n8pb7\" (UID: \"b414a389-9e48-412c-b4e0-26a9b4cb51eb\") " pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:08:51 crc kubenswrapper[4646]: I1203 11:08:51.137555 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b414a389-9e48-412c-b4e0-26a9b4cb51eb-catalog-content\") pod \"community-operators-n8pb7\" (UID: \"b414a389-9e48-412c-b4e0-26a9b4cb51eb\") " pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:08:51 crc kubenswrapper[4646]: I1203 11:08:51.137951 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b414a389-9e48-412c-b4e0-26a9b4cb51eb-catalog-content\") pod \"community-operators-n8pb7\" (UID: \"b414a389-9e48-412c-b4e0-26a9b4cb51eb\") " pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:08:51 crc kubenswrapper[4646]: I1203 11:08:51.138006 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b414a389-9e48-412c-b4e0-26a9b4cb51eb-utilities\") pod \"community-operators-n8pb7\" (UID: \"b414a389-9e48-412c-b4e0-26a9b4cb51eb\") " pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:08:51 crc kubenswrapper[4646]: I1203 11:08:51.138424 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b414a389-9e48-412c-b4e0-26a9b4cb51eb-utilities\") pod \"community-operators-n8pb7\" (UID: \"b414a389-9e48-412c-b4e0-26a9b4cb51eb\") " pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:08:51 crc kubenswrapper[4646]: I1203 11:08:51.138574 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjz45\" (UniqueName: \"kubernetes.io/projected/b414a389-9e48-412c-b4e0-26a9b4cb51eb-kube-api-access-hjz45\") pod \"community-operators-n8pb7\" (UID: \"b414a389-9e48-412c-b4e0-26a9b4cb51eb\") " pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:08:51 crc kubenswrapper[4646]: I1203 11:08:51.162692 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjz45\" (UniqueName: \"kubernetes.io/projected/b414a389-9e48-412c-b4e0-26a9b4cb51eb-kube-api-access-hjz45\") pod \"community-operators-n8pb7\" (UID: \"b414a389-9e48-412c-b4e0-26a9b4cb51eb\") " pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:08:51 crc kubenswrapper[4646]: I1203 11:08:51.361193 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:08:51 crc kubenswrapper[4646]: I1203 11:08:51.941736 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n8pb7"] Dec 03 11:08:53 crc kubenswrapper[4646]: I1203 11:08:53.868768 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hblfs"] Dec 03 11:08:53 crc kubenswrapper[4646]: I1203 11:08:53.871772 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:08:53 crc kubenswrapper[4646]: I1203 11:08:53.881658 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hblfs"] Dec 03 11:08:53 crc kubenswrapper[4646]: I1203 11:08:53.927093 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42a46b9b-c088-4e78-848e-c5b605694d3e-catalog-content\") pod \"redhat-marketplace-hblfs\" (UID: \"42a46b9b-c088-4e78-848e-c5b605694d3e\") " pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:08:53 crc kubenswrapper[4646]: I1203 11:08:53.927170 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42a46b9b-c088-4e78-848e-c5b605694d3e-utilities\") pod \"redhat-marketplace-hblfs\" (UID: \"42a46b9b-c088-4e78-848e-c5b605694d3e\") " pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:08:53 crc kubenswrapper[4646]: I1203 11:08:53.927241 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgh6r\" (UniqueName: \"kubernetes.io/projected/42a46b9b-c088-4e78-848e-c5b605694d3e-kube-api-access-lgh6r\") pod \"redhat-marketplace-hblfs\" (UID: \"42a46b9b-c088-4e78-848e-c5b605694d3e\") " pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:08:54 crc kubenswrapper[4646]: I1203 11:08:54.028453 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgh6r\" (UniqueName: \"kubernetes.io/projected/42a46b9b-c088-4e78-848e-c5b605694d3e-kube-api-access-lgh6r\") pod \"redhat-marketplace-hblfs\" (UID: \"42a46b9b-c088-4e78-848e-c5b605694d3e\") " pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:08:54 crc kubenswrapper[4646]: I1203 11:08:54.028527 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42a46b9b-c088-4e78-848e-c5b605694d3e-catalog-content\") pod \"redhat-marketplace-hblfs\" (UID: \"42a46b9b-c088-4e78-848e-c5b605694d3e\") " pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:08:54 crc kubenswrapper[4646]: I1203 11:08:54.028547 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42a46b9b-c088-4e78-848e-c5b605694d3e-utilities\") pod \"redhat-marketplace-hblfs\" (UID: \"42a46b9b-c088-4e78-848e-c5b605694d3e\") " pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:08:54 crc kubenswrapper[4646]: I1203 11:08:54.029045 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42a46b9b-c088-4e78-848e-c5b605694d3e-catalog-content\") pod \"redhat-marketplace-hblfs\" (UID: \"42a46b9b-c088-4e78-848e-c5b605694d3e\") " pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:08:54 crc kubenswrapper[4646]: I1203 11:08:54.029074 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42a46b9b-c088-4e78-848e-c5b605694d3e-utilities\") pod \"redhat-marketplace-hblfs\" (UID: \"42a46b9b-c088-4e78-848e-c5b605694d3e\") " pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:08:54 crc kubenswrapper[4646]: I1203 11:08:54.048918 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgh6r\" (UniqueName: \"kubernetes.io/projected/42a46b9b-c088-4e78-848e-c5b605694d3e-kube-api-access-lgh6r\") pod \"redhat-marketplace-hblfs\" (UID: \"42a46b9b-c088-4e78-848e-c5b605694d3e\") " pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:08:54 crc kubenswrapper[4646]: I1203 11:08:54.206871 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.426421 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.427856 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.429518 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-frrsv" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.438486 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.439723 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.440961 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-2vsbp" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.446524 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.463025 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.469944 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.471024 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.475234 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-pk6kh" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.496391 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.497684 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.505798 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.507678 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-8mmmr" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.534950 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.535686 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnmkm\" (UniqueName: \"kubernetes.io/projected/4cb5b92a-540a-41d9-a662-b1f3faf07829-kube-api-access-dnmkm\") pod \"designate-operator-controller-manager-78b4bc895b-ps2zm\" (UID: \"4cb5b92a-540a-41d9-a662-b1f3faf07829\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.535733 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmpxh\" (UniqueName: \"kubernetes.io/projected/e400aa20-0649-47de-8f10-da6090800eb3-kube-api-access-tmpxh\") pod \"barbican-operator-controller-manager-7d9dfd778-ss5tr\" (UID: \"e400aa20-0649-47de-8f10-da6090800eb3\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.535785 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kmmc\" (UniqueName: \"kubernetes.io/projected/d7ed2042-fe69-40f6-8440-49d63e851ffc-kube-api-access-5kmmc\") pod \"cinder-operator-controller-manager-859b6ccc6-pnbk8\" (UID: \"d7ed2042-fe69-40f6-8440-49d63e851ffc\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.539324 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mzlrq"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.540525 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mzlrq" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.544469 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-lb4bx" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.551528 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mzlrq"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.577094 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6hl22"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.582656 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6hl22" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.589279 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-z524h" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.614651 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6hl22"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.636831 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnmkm\" (UniqueName: \"kubernetes.io/projected/4cb5b92a-540a-41d9-a662-b1f3faf07829-kube-api-access-dnmkm\") pod \"designate-operator-controller-manager-78b4bc895b-ps2zm\" (UID: \"4cb5b92a-540a-41d9-a662-b1f3faf07829\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.637125 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmpxh\" (UniqueName: \"kubernetes.io/projected/e400aa20-0649-47de-8f10-da6090800eb3-kube-api-access-tmpxh\") pod \"barbican-operator-controller-manager-7d9dfd778-ss5tr\" (UID: \"e400aa20-0649-47de-8f10-da6090800eb3\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.637320 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kmmc\" (UniqueName: \"kubernetes.io/projected/d7ed2042-fe69-40f6-8440-49d63e851ffc-kube-api-access-5kmmc\") pod \"cinder-operator-controller-manager-859b6ccc6-pnbk8\" (UID: \"d7ed2042-fe69-40f6-8440-49d63e851ffc\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.637457 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st64w\" (UniqueName: \"kubernetes.io/projected/396bd18c-37f7-471f-a0f9-0b9cb5a447e3-kube-api-access-st64w\") pod \"heat-operator-controller-manager-5f64f6f8bb-mzlrq\" (UID: \"396bd18c-37f7-471f-a0f9-0b9cb5a447e3\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mzlrq" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.637584 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r764p\" (UniqueName: \"kubernetes.io/projected/1fed03a1-2388-470c-843e-4de04fd6d9bc-kube-api-access-r764p\") pod \"glance-operator-controller-manager-77987cd8cd-56r4n\" (UID: \"1fed03a1-2388-470c-843e-4de04fd6d9bc\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.637726 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4qxc\" (UniqueName: \"kubernetes.io/projected/089f6f35-ab06-4b31-a331-16411a9783e7-kube-api-access-p4qxc\") pod \"horizon-operator-controller-manager-68c6d99b8f-6hl22\" (UID: \"089f6f35-ab06-4b31-a331-16411a9783e7\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6hl22" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.641597 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.642746 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.647933 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.651565 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-4pnkr" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.676676 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.689258 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-schbl"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.691047 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-schbl" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.700082 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kmmc\" (UniqueName: \"kubernetes.io/projected/d7ed2042-fe69-40f6-8440-49d63e851ffc-kube-api-access-5kmmc\") pod \"cinder-operator-controller-manager-859b6ccc6-pnbk8\" (UID: \"d7ed2042-fe69-40f6-8440-49d63e851ffc\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.708870 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmpxh\" (UniqueName: \"kubernetes.io/projected/e400aa20-0649-47de-8f10-da6090800eb3-kube-api-access-tmpxh\") pod \"barbican-operator-controller-manager-7d9dfd778-ss5tr\" (UID: \"e400aa20-0649-47de-8f10-da6090800eb3\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.710046 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnmkm\" (UniqueName: \"kubernetes.io/projected/4cb5b92a-540a-41d9-a662-b1f3faf07829-kube-api-access-dnmkm\") pod \"designate-operator-controller-manager-78b4bc895b-ps2zm\" (UID: \"4cb5b92a-540a-41d9-a662-b1f3faf07829\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.710346 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-7zj2g" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.727387 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-l7v6k"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.728815 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-l7v6k" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.737812 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-nkqwb" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.738616 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st64w\" (UniqueName: \"kubernetes.io/projected/396bd18c-37f7-471f-a0f9-0b9cb5a447e3-kube-api-access-st64w\") pod \"heat-operator-controller-manager-5f64f6f8bb-mzlrq\" (UID: \"396bd18c-37f7-471f-a0f9-0b9cb5a447e3\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mzlrq" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.738667 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r764p\" (UniqueName: \"kubernetes.io/projected/1fed03a1-2388-470c-843e-4de04fd6d9bc-kube-api-access-r764p\") pod \"glance-operator-controller-manager-77987cd8cd-56r4n\" (UID: \"1fed03a1-2388-470c-843e-4de04fd6d9bc\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.738701 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert\") pod \"infra-operator-controller-manager-57548d458d-g6jbp\" (UID: \"d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.738750 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxq2c\" (UniqueName: \"kubernetes.io/projected/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-kube-api-access-kxq2c\") pod \"infra-operator-controller-manager-57548d458d-g6jbp\" (UID: \"d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.738786 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npqn5\" (UniqueName: \"kubernetes.io/projected/31371552-226a-436f-a595-52bc387ed5cf-kube-api-access-npqn5\") pod \"ironic-operator-controller-manager-6c548fd776-schbl\" (UID: \"31371552-226a-436f-a595-52bc387ed5cf\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-schbl" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.738814 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4qxc\" (UniqueName: \"kubernetes.io/projected/089f6f35-ab06-4b31-a331-16411a9783e7-kube-api-access-p4qxc\") pod \"horizon-operator-controller-manager-68c6d99b8f-6hl22\" (UID: \"089f6f35-ab06-4b31-a331-16411a9783e7\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6hl22" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.751635 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-l7v6k"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.755713 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.759466 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-rcmlq"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.760477 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rcmlq" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.763318 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r764p\" (UniqueName: \"kubernetes.io/projected/1fed03a1-2388-470c-843e-4de04fd6d9bc-kube-api-access-r764p\") pod \"glance-operator-controller-manager-77987cd8cd-56r4n\" (UID: \"1fed03a1-2388-470c-843e-4de04fd6d9bc\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.763853 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4qxc\" (UniqueName: \"kubernetes.io/projected/089f6f35-ab06-4b31-a331-16411a9783e7-kube-api-access-p4qxc\") pod \"horizon-operator-controller-manager-68c6d99b8f-6hl22\" (UID: \"089f6f35-ab06-4b31-a331-16411a9783e7\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6hl22" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.767932 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.768060 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-mmxqc" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.773522 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-schbl"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.776925 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-rcmlq"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.784000 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st64w\" (UniqueName: \"kubernetes.io/projected/396bd18c-37f7-471f-a0f9-0b9cb5a447e3-kube-api-access-st64w\") pod \"heat-operator-controller-manager-5f64f6f8bb-mzlrq\" (UID: \"396bd18c-37f7-471f-a0f9-0b9cb5a447e3\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mzlrq" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.788644 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.796084 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-92vgw"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.797726 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-92vgw" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.808686 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-68b48" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.813215 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.827402 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.828812 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.833387 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-rljnm" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.839448 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxq2c\" (UniqueName: \"kubernetes.io/projected/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-kube-api-access-kxq2c\") pod \"infra-operator-controller-manager-57548d458d-g6jbp\" (UID: \"d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.839634 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npqn5\" (UniqueName: \"kubernetes.io/projected/31371552-226a-436f-a595-52bc387ed5cf-kube-api-access-npqn5\") pod \"ironic-operator-controller-manager-6c548fd776-schbl\" (UID: \"31371552-226a-436f-a595-52bc387ed5cf\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-schbl" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.839763 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnf4m\" (UniqueName: \"kubernetes.io/projected/68e6047b-627f-4c45-aa55-4aeab90e17b5-kube-api-access-gnf4m\") pod \"keystone-operator-controller-manager-7765d96ddf-l7v6k\" (UID: \"68e6047b-627f-4c45-aa55-4aeab90e17b5\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-l7v6k" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.839874 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmnwd\" (UniqueName: \"kubernetes.io/projected/f3e07f77-c87f-4c6b-be4e-938d71354f8b-kube-api-access-wmnwd\") pod \"manila-operator-controller-manager-7c79b5df47-rcmlq\" (UID: \"f3e07f77-c87f-4c6b-be4e-938d71354f8b\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rcmlq" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.839971 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wd4d\" (UniqueName: \"kubernetes.io/projected/a0fa06f6-534e-4ad4-806d-c13cad106809-kube-api-access-4wd4d\") pod \"mariadb-operator-controller-manager-56bbcc9d85-92vgw\" (UID: \"a0fa06f6-534e-4ad4-806d-c13cad106809\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-92vgw" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.840072 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert\") pod \"infra-operator-controller-manager-57548d458d-g6jbp\" (UID: \"d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" Dec 03 11:08:57 crc kubenswrapper[4646]: E1203 11:08:57.840307 4646 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 11:08:57 crc kubenswrapper[4646]: E1203 11:08:57.840495 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert podName:d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0 nodeName:}" failed. No retries permitted until 2025-12-03 11:08:58.340477936 +0000 UTC m=+914.803534061 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert") pod "infra-operator-controller-manager-57548d458d-g6jbp" (UID: "d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0") : secret "infra-operator-webhook-server-cert" not found Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.845405 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-92vgw"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.854720 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mzlrq" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.870108 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxq2c\" (UniqueName: \"kubernetes.io/projected/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-kube-api-access-kxq2c\") pod \"infra-operator-controller-manager-57548d458d-g6jbp\" (UID: \"d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.887089 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sc2vn"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.887980 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.888064 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sc2vn" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.890888 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-lg52r" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.894235 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npqn5\" (UniqueName: \"kubernetes.io/projected/31371552-226a-436f-a595-52bc387ed5cf-kube-api-access-npqn5\") pod \"ironic-operator-controller-manager-6c548fd776-schbl\" (UID: \"31371552-226a-436f-a595-52bc387ed5cf\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-schbl" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.896917 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-n9s99"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.898071 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.902215 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sc2vn"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.911670 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6hl22" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.917645 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-n9s99"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.919041 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-8rccw" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.930820 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.931999 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.936685 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-4xhq7" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.940225 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.940878 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmnwd\" (UniqueName: \"kubernetes.io/projected/f3e07f77-c87f-4c6b-be4e-938d71354f8b-kube-api-access-wmnwd\") pod \"manila-operator-controller-manager-7c79b5df47-rcmlq\" (UID: \"f3e07f77-c87f-4c6b-be4e-938d71354f8b\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rcmlq" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.940950 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wd4d\" (UniqueName: \"kubernetes.io/projected/a0fa06f6-534e-4ad4-806d-c13cad106809-kube-api-access-4wd4d\") pod \"mariadb-operator-controller-manager-56bbcc9d85-92vgw\" (UID: \"a0fa06f6-534e-4ad4-806d-c13cad106809\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-92vgw" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.941066 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4bk6\" (UniqueName: \"kubernetes.io/projected/9a1416d1-243f-4ab9-9a9c-4b71e0f23876-kube-api-access-j4bk6\") pod \"nova-operator-controller-manager-697bc559fc-g8pg8\" (UID: \"9a1416d1-243f-4ab9-9a9c-4b71e0f23876\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.941093 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnf4m\" (UniqueName: \"kubernetes.io/projected/68e6047b-627f-4c45-aa55-4aeab90e17b5-kube-api-access-gnf4m\") pod \"keystone-operator-controller-manager-7765d96ddf-l7v6k\" (UID: \"68e6047b-627f-4c45-aa55-4aeab90e17b5\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-l7v6k" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.941111 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs8bq\" (UniqueName: \"kubernetes.io/projected/ed761af7-289e-4f6c-a7b9-710886250a3a-kube-api-access-fs8bq\") pod \"octavia-operator-controller-manager-998648c74-n9s99\" (UID: \"ed761af7-289e-4f6c-a7b9-710886250a3a\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.941145 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qjsn\" (UniqueName: \"kubernetes.io/projected/eaefadf5-75cd-4ff2-811c-e25bb2850173-kube-api-access-9qjsn\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-sc2vn\" (UID: \"eaefadf5-75cd-4ff2-811c-e25bb2850173\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sc2vn" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.948376 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.954649 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-t7fj8"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.957924 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-t7fj8" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.964933 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-c26gb" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.982426 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-f7pcq"] Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.998220 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnf4m\" (UniqueName: \"kubernetes.io/projected/68e6047b-627f-4c45-aa55-4aeab90e17b5-kube-api-access-gnf4m\") pod \"keystone-operator-controller-manager-7765d96ddf-l7v6k\" (UID: \"68e6047b-627f-4c45-aa55-4aeab90e17b5\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-l7v6k" Dec 03 11:08:57 crc kubenswrapper[4646]: I1203 11:08:57.998988 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmnwd\" (UniqueName: \"kubernetes.io/projected/f3e07f77-c87f-4c6b-be4e-938d71354f8b-kube-api-access-wmnwd\") pod \"manila-operator-controller-manager-7c79b5df47-rcmlq\" (UID: \"f3e07f77-c87f-4c6b-be4e-938d71354f8b\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rcmlq" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.002850 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wd4d\" (UniqueName: \"kubernetes.io/projected/a0fa06f6-534e-4ad4-806d-c13cad106809-kube-api-access-4wd4d\") pod \"mariadb-operator-controller-manager-56bbcc9d85-92vgw\" (UID: \"a0fa06f6-534e-4ad4-806d-c13cad106809\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-92vgw" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.044162 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4bk6\" (UniqueName: \"kubernetes.io/projected/9a1416d1-243f-4ab9-9a9c-4b71e0f23876-kube-api-access-j4bk6\") pod \"nova-operator-controller-manager-697bc559fc-g8pg8\" (UID: \"9a1416d1-243f-4ab9-9a9c-4b71e0f23876\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.044216 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs8bq\" (UniqueName: \"kubernetes.io/projected/ed761af7-289e-4f6c-a7b9-710886250a3a-kube-api-access-fs8bq\") pod \"octavia-operator-controller-manager-998648c74-n9s99\" (UID: \"ed761af7-289e-4f6c-a7b9-710886250a3a\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.044258 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qjsn\" (UniqueName: \"kubernetes.io/projected/eaefadf5-75cd-4ff2-811c-e25bb2850173-kube-api-access-9qjsn\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-sc2vn\" (UID: \"eaefadf5-75cd-4ff2-811c-e25bb2850173\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sc2vn" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.044291 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9\" (UID: \"e776d6da-584d-498f-b611-0eb67be6d6d0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.044324 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77857\" (UniqueName: \"kubernetes.io/projected/e776d6da-584d-498f-b611-0eb67be6d6d0-kube-api-access-77857\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9\" (UID: \"e776d6da-584d-498f-b611-0eb67be6d6d0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.044392 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47k8t\" (UniqueName: \"kubernetes.io/projected/153f2bfa-6ca2-4a9f-900e-5321c5791639-kube-api-access-47k8t\") pod \"ovn-operator-controller-manager-b6456fdb6-t7fj8\" (UID: \"153f2bfa-6ca2-4a9f-900e-5321c5791639\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-t7fj8" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.055380 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-f7pcq" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.067823 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-t7fj8"] Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.070573 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-schbl" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.080081 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-bnhtb" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.084421 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-f7pcq"] Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.090748 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs8bq\" (UniqueName: \"kubernetes.io/projected/ed761af7-289e-4f6c-a7b9-710886250a3a-kube-api-access-fs8bq\") pod \"octavia-operator-controller-manager-998648c74-n9s99\" (UID: \"ed761af7-289e-4f6c-a7b9-710886250a3a\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.103942 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-j2bwb"] Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.106776 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-j2bwb" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.124320 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qjsn\" (UniqueName: \"kubernetes.io/projected/eaefadf5-75cd-4ff2-811c-e25bb2850173-kube-api-access-9qjsn\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-sc2vn\" (UID: \"eaefadf5-75cd-4ff2-811c-e25bb2850173\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sc2vn" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.128876 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-j2bwb"] Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.134311 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-l7v6k" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.140196 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4bk6\" (UniqueName: \"kubernetes.io/projected/9a1416d1-243f-4ab9-9a9c-4b71e0f23876-kube-api-access-j4bk6\") pod \"nova-operator-controller-manager-697bc559fc-g8pg8\" (UID: \"9a1416d1-243f-4ab9-9a9c-4b71e0f23876\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.140533 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-wflj6" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.141135 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-w5c45"] Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.142547 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rcmlq" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.143303 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-w5c45" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.145191 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-pgllf" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.146209 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9\" (UID: \"e776d6da-584d-498f-b611-0eb67be6d6d0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.146240 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt528\" (UniqueName: \"kubernetes.io/projected/011cd55c-5e76-4dd3-905b-62c587119096-kube-api-access-mt528\") pod \"placement-operator-controller-manager-78f8948974-f7pcq\" (UID: \"011cd55c-5e76-4dd3-905b-62c587119096\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-f7pcq" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.146269 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77857\" (UniqueName: \"kubernetes.io/projected/e776d6da-584d-498f-b611-0eb67be6d6d0-kube-api-access-77857\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9\" (UID: \"e776d6da-584d-498f-b611-0eb67be6d6d0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.146295 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47k8t\" (UniqueName: \"kubernetes.io/projected/153f2bfa-6ca2-4a9f-900e-5321c5791639-kube-api-access-47k8t\") pod \"ovn-operator-controller-manager-b6456fdb6-t7fj8\" (UID: \"153f2bfa-6ca2-4a9f-900e-5321c5791639\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-t7fj8" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.146345 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nsls\" (UniqueName: \"kubernetes.io/projected/f274e7e0-a406-4617-95bc-fbca7ef39526-kube-api-access-2nsls\") pod \"swift-operator-controller-manager-5f8c65bbfc-j2bwb\" (UID: \"f274e7e0-a406-4617-95bc-fbca7ef39526\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-j2bwb" Dec 03 11:08:58 crc kubenswrapper[4646]: E1203 11:08:58.146494 4646 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 11:08:58 crc kubenswrapper[4646]: E1203 11:08:58.146534 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert podName:e776d6da-584d-498f-b611-0eb67be6d6d0 nodeName:}" failed. No retries permitted until 2025-12-03 11:08:58.646519815 +0000 UTC m=+915.109575950 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" (UID: "e776d6da-584d-498f-b611-0eb67be6d6d0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.157541 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-z6m86"] Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.159097 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6m86" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.161900 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-92vgw" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.172563 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-hlwj7"] Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.174089 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-w5c45"] Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.174178 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hlwj7" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.177969 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-z6m86"] Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.182442 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-hlwj7"] Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.183793 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-vld9x" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.188995 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-wjzxw" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.198081 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77857\" (UniqueName: \"kubernetes.io/projected/e776d6da-584d-498f-b611-0eb67be6d6d0-kube-api-access-77857\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9\" (UID: \"e776d6da-584d-498f-b611-0eb67be6d6d0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.206997 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47k8t\" (UniqueName: \"kubernetes.io/projected/153f2bfa-6ca2-4a9f-900e-5321c5791639-kube-api-access-47k8t\") pod \"ovn-operator-controller-manager-b6456fdb6-t7fj8\" (UID: \"153f2bfa-6ca2-4a9f-900e-5321c5791639\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-t7fj8" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.224394 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.251691 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sc2vn" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.252156 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nsls\" (UniqueName: \"kubernetes.io/projected/f274e7e0-a406-4617-95bc-fbca7ef39526-kube-api-access-2nsls\") pod \"swift-operator-controller-manager-5f8c65bbfc-j2bwb\" (UID: \"f274e7e0-a406-4617-95bc-fbca7ef39526\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-j2bwb" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.252395 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6g6n\" (UniqueName: \"kubernetes.io/projected/55e5ed61-839c-441b-acf2-cd43cc0c5c90-kube-api-access-h6g6n\") pod \"telemetry-operator-controller-manager-76cc84c6bb-w5c45\" (UID: \"55e5ed61-839c-441b-acf2-cd43cc0c5c90\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-w5c45" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.252450 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr459\" (UniqueName: \"kubernetes.io/projected/0be433f8-aae1-48a4-8853-de55ae48f836-kube-api-access-tr459\") pod \"test-operator-controller-manager-5854674fcc-z6m86\" (UID: \"0be433f8-aae1-48a4-8853-de55ae48f836\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6m86" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.252533 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt528\" (UniqueName: \"kubernetes.io/projected/011cd55c-5e76-4dd3-905b-62c587119096-kube-api-access-mt528\") pod \"placement-operator-controller-manager-78f8948974-f7pcq\" (UID: \"011cd55c-5e76-4dd3-905b-62c587119096\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-f7pcq" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.252601 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prm85\" (UniqueName: \"kubernetes.io/projected/e2fc2280-6814-4ba0-bd48-2272d37f2b41-kube-api-access-prm85\") pod \"watcher-operator-controller-manager-769dc69bc-hlwj7\" (UID: \"e2fc2280-6814-4ba0-bd48-2272d37f2b41\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hlwj7" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.274708 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.291986 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nsls\" (UniqueName: \"kubernetes.io/projected/f274e7e0-a406-4617-95bc-fbca7ef39526-kube-api-access-2nsls\") pod \"swift-operator-controller-manager-5f8c65bbfc-j2bwb\" (UID: \"f274e7e0-a406-4617-95bc-fbca7ef39526\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-j2bwb" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.298011 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt528\" (UniqueName: \"kubernetes.io/projected/011cd55c-5e76-4dd3-905b-62c587119096-kube-api-access-mt528\") pod \"placement-operator-controller-manager-78f8948974-f7pcq\" (UID: \"011cd55c-5e76-4dd3-905b-62c587119096\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-f7pcq" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.319570 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-t7fj8" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.337101 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7"] Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.342612 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.345708 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-z667w" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.345867 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.345964 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.354943 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prm85\" (UniqueName: \"kubernetes.io/projected/e2fc2280-6814-4ba0-bd48-2272d37f2b41-kube-api-access-prm85\") pod \"watcher-operator-controller-manager-769dc69bc-hlwj7\" (UID: \"e2fc2280-6814-4ba0-bd48-2272d37f2b41\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hlwj7" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.354983 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.355027 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert\") pod \"infra-operator-controller-manager-57548d458d-g6jbp\" (UID: \"d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.355048 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.355091 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zclql\" (UniqueName: \"kubernetes.io/projected/83f119ed-6576-4e47-82b1-a06189cff0a7-kube-api-access-zclql\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.355123 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6g6n\" (UniqueName: \"kubernetes.io/projected/55e5ed61-839c-441b-acf2-cd43cc0c5c90-kube-api-access-h6g6n\") pod \"telemetry-operator-controller-manager-76cc84c6bb-w5c45\" (UID: \"55e5ed61-839c-441b-acf2-cd43cc0c5c90\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-w5c45" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.355140 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr459\" (UniqueName: \"kubernetes.io/projected/0be433f8-aae1-48a4-8853-de55ae48f836-kube-api-access-tr459\") pod \"test-operator-controller-manager-5854674fcc-z6m86\" (UID: \"0be433f8-aae1-48a4-8853-de55ae48f836\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6m86" Dec 03 11:08:58 crc kubenswrapper[4646]: E1203 11:08:58.355395 4646 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 11:08:58 crc kubenswrapper[4646]: E1203 11:08:58.355470 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert podName:d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0 nodeName:}" failed. No retries permitted until 2025-12-03 11:08:59.355447546 +0000 UTC m=+915.818503751 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert") pod "infra-operator-controller-manager-57548d458d-g6jbp" (UID: "d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0") : secret "infra-operator-webhook-server-cert" not found Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.374429 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7"] Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.386040 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr459\" (UniqueName: \"kubernetes.io/projected/0be433f8-aae1-48a4-8853-de55ae48f836-kube-api-access-tr459\") pod \"test-operator-controller-manager-5854674fcc-z6m86\" (UID: \"0be433f8-aae1-48a4-8853-de55ae48f836\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6m86" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.405806 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prm85\" (UniqueName: \"kubernetes.io/projected/e2fc2280-6814-4ba0-bd48-2272d37f2b41-kube-api-access-prm85\") pod \"watcher-operator-controller-manager-769dc69bc-hlwj7\" (UID: \"e2fc2280-6814-4ba0-bd48-2272d37f2b41\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hlwj7" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.421482 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6g6n\" (UniqueName: \"kubernetes.io/projected/55e5ed61-839c-441b-acf2-cd43cc0c5c90-kube-api-access-h6g6n\") pod \"telemetry-operator-controller-manager-76cc84c6bb-w5c45\" (UID: \"55e5ed61-839c-441b-acf2-cd43cc0c5c90\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-w5c45" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.435675 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-f7pcq" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.440590 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-j2bwb" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.456113 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.456179 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.456234 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zclql\" (UniqueName: \"kubernetes.io/projected/83f119ed-6576-4e47-82b1-a06189cff0a7-kube-api-access-zclql\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:08:58 crc kubenswrapper[4646]: E1203 11:08:58.456608 4646 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 11:08:58 crc kubenswrapper[4646]: E1203 11:08:58.456655 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs podName:83f119ed-6576-4e47-82b1-a06189cff0a7 nodeName:}" failed. No retries permitted until 2025-12-03 11:08:58.956640552 +0000 UTC m=+915.419696687 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs") pod "openstack-operator-controller-manager-6c5c989645-kkkb7" (UID: "83f119ed-6576-4e47-82b1-a06189cff0a7") : secret "webhook-server-cert" not found Dec 03 11:08:58 crc kubenswrapper[4646]: E1203 11:08:58.456811 4646 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 11:08:58 crc kubenswrapper[4646]: E1203 11:08:58.456840 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs podName:83f119ed-6576-4e47-82b1-a06189cff0a7 nodeName:}" failed. No retries permitted until 2025-12-03 11:08:58.956831617 +0000 UTC m=+915.419887752 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs") pod "openstack-operator-controller-manager-6c5c989645-kkkb7" (UID: "83f119ed-6576-4e47-82b1-a06189cff0a7") : secret "metrics-server-cert" not found Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.482186 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zclql\" (UniqueName: \"kubernetes.io/projected/83f119ed-6576-4e47-82b1-a06189cff0a7-kube-api-access-zclql\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.499688 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5vdh"] Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.500486 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5vdh" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.506868 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-z6tx2" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.517800 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5vdh"] Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.518060 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-w5c45" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.545622 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6m86" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.552032 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hlwj7" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.557577 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfvsz\" (UniqueName: \"kubernetes.io/projected/afd2b4e9-59a4-4e52-bd72-706b8edbd719-kube-api-access-cfvsz\") pod \"rabbitmq-cluster-operator-manager-668c99d594-b5vdh\" (UID: \"afd2b4e9-59a4-4e52-bd72-706b8edbd719\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5vdh" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.658907 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfvsz\" (UniqueName: \"kubernetes.io/projected/afd2b4e9-59a4-4e52-bd72-706b8edbd719-kube-api-access-cfvsz\") pod \"rabbitmq-cluster-operator-manager-668c99d594-b5vdh\" (UID: \"afd2b4e9-59a4-4e52-bd72-706b8edbd719\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5vdh" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.658962 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9\" (UID: \"e776d6da-584d-498f-b611-0eb67be6d6d0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" Dec 03 11:08:58 crc kubenswrapper[4646]: E1203 11:08:58.659094 4646 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 11:08:58 crc kubenswrapper[4646]: E1203 11:08:58.659138 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert podName:e776d6da-584d-498f-b611-0eb67be6d6d0 nodeName:}" failed. No retries permitted until 2025-12-03 11:08:59.659122416 +0000 UTC m=+916.122178551 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" (UID: "e776d6da-584d-498f-b611-0eb67be6d6d0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.674368 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfvsz\" (UniqueName: \"kubernetes.io/projected/afd2b4e9-59a4-4e52-bd72-706b8edbd719-kube-api-access-cfvsz\") pod \"rabbitmq-cluster-operator-manager-668c99d594-b5vdh\" (UID: \"afd2b4e9-59a4-4e52-bd72-706b8edbd719\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5vdh" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.812829 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5vdh" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.963012 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:08:58 crc kubenswrapper[4646]: I1203 11:08:58.963137 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:08:58 crc kubenswrapper[4646]: E1203 11:08:58.963208 4646 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 11:08:58 crc kubenswrapper[4646]: E1203 11:08:58.963252 4646 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 11:08:58 crc kubenswrapper[4646]: E1203 11:08:58.963299 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs podName:83f119ed-6576-4e47-82b1-a06189cff0a7 nodeName:}" failed. No retries permitted until 2025-12-03 11:08:59.96328054 +0000 UTC m=+916.426336675 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs") pod "openstack-operator-controller-manager-6c5c989645-kkkb7" (UID: "83f119ed-6576-4e47-82b1-a06189cff0a7") : secret "metrics-server-cert" not found Dec 03 11:08:58 crc kubenswrapper[4646]: E1203 11:08:58.963768 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs podName:83f119ed-6576-4e47-82b1-a06189cff0a7 nodeName:}" failed. No retries permitted until 2025-12-03 11:08:59.963310261 +0000 UTC m=+916.426366496 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs") pod "openstack-operator-controller-manager-6c5c989645-kkkb7" (UID: "83f119ed-6576-4e47-82b1-a06189cff0a7") : secret "webhook-server-cert" not found Dec 03 11:08:59 crc kubenswrapper[4646]: I1203 11:08:59.367377 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert\") pod \"infra-operator-controller-manager-57548d458d-g6jbp\" (UID: \"d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" Dec 03 11:08:59 crc kubenswrapper[4646]: E1203 11:08:59.367589 4646 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 11:08:59 crc kubenswrapper[4646]: E1203 11:08:59.367638 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert podName:d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0 nodeName:}" failed. No retries permitted until 2025-12-03 11:09:01.367624501 +0000 UTC m=+917.830680636 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert") pod "infra-operator-controller-manager-57548d458d-g6jbp" (UID: "d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0") : secret "infra-operator-webhook-server-cert" not found Dec 03 11:08:59 crc kubenswrapper[4646]: I1203 11:08:59.670565 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9\" (UID: \"e776d6da-584d-498f-b611-0eb67be6d6d0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" Dec 03 11:08:59 crc kubenswrapper[4646]: E1203 11:08:59.670742 4646 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 11:08:59 crc kubenswrapper[4646]: E1203 11:08:59.670809 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert podName:e776d6da-584d-498f-b611-0eb67be6d6d0 nodeName:}" failed. No retries permitted until 2025-12-03 11:09:01.670788876 +0000 UTC m=+918.133845011 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" (UID: "e776d6da-584d-498f-b611-0eb67be6d6d0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 11:08:59 crc kubenswrapper[4646]: I1203 11:08:59.980716 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:08:59 crc kubenswrapper[4646]: I1203 11:08:59.980781 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:08:59 crc kubenswrapper[4646]: E1203 11:08:59.980844 4646 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 11:08:59 crc kubenswrapper[4646]: E1203 11:08:59.980867 4646 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 11:08:59 crc kubenswrapper[4646]: E1203 11:08:59.980897 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs podName:83f119ed-6576-4e47-82b1-a06189cff0a7 nodeName:}" failed. No retries permitted until 2025-12-03 11:09:01.980881532 +0000 UTC m=+918.443937667 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs") pod "openstack-operator-controller-manager-6c5c989645-kkkb7" (UID: "83f119ed-6576-4e47-82b1-a06189cff0a7") : secret "webhook-server-cert" not found Dec 03 11:08:59 crc kubenswrapper[4646]: E1203 11:08:59.980910 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs podName:83f119ed-6576-4e47-82b1-a06189cff0a7 nodeName:}" failed. No retries permitted until 2025-12-03 11:09:01.980905282 +0000 UTC m=+918.443961407 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs") pod "openstack-operator-controller-manager-6c5c989645-kkkb7" (UID: "83f119ed-6576-4e47-82b1-a06189cff0a7") : secret "metrics-server-cert" not found Dec 03 11:09:01 crc kubenswrapper[4646]: I1203 11:09:01.181318 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8pb7" event={"ID":"b414a389-9e48-412c-b4e0-26a9b4cb51eb","Type":"ContainerStarted","Data":"5552dde4963815fc95260e1b3fa27fa5397491c451448fc32268c147176942aa"} Dec 03 11:09:01 crc kubenswrapper[4646]: I1203 11:09:01.398039 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert\") pod \"infra-operator-controller-manager-57548d458d-g6jbp\" (UID: \"d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" Dec 03 11:09:01 crc kubenswrapper[4646]: E1203 11:09:01.398223 4646 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 11:09:01 crc kubenswrapper[4646]: E1203 11:09:01.398268 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert podName:d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0 nodeName:}" failed. No retries permitted until 2025-12-03 11:09:05.398255881 +0000 UTC m=+921.861312016 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert") pod "infra-operator-controller-manager-57548d458d-g6jbp" (UID: "d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0") : secret "infra-operator-webhook-server-cert" not found Dec 03 11:09:01 crc kubenswrapper[4646]: I1203 11:09:01.719791 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9\" (UID: \"e776d6da-584d-498f-b611-0eb67be6d6d0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" Dec 03 11:09:01 crc kubenswrapper[4646]: E1203 11:09:01.719928 4646 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 11:09:01 crc kubenswrapper[4646]: E1203 11:09:01.719972 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert podName:e776d6da-584d-498f-b611-0eb67be6d6d0 nodeName:}" failed. No retries permitted until 2025-12-03 11:09:05.719957552 +0000 UTC m=+922.183013687 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" (UID: "e776d6da-584d-498f-b611-0eb67be6d6d0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 11:09:01 crc kubenswrapper[4646]: I1203 11:09:01.743208 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hblfs"] Dec 03 11:09:02 crc kubenswrapper[4646]: I1203 11:09:02.035084 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:09:02 crc kubenswrapper[4646]: I1203 11:09:02.035380 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:09:02 crc kubenswrapper[4646]: E1203 11:09:02.035753 4646 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 11:09:02 crc kubenswrapper[4646]: E1203 11:09:02.035801 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs podName:83f119ed-6576-4e47-82b1-a06189cff0a7 nodeName:}" failed. No retries permitted until 2025-12-03 11:09:06.035787844 +0000 UTC m=+922.498843979 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs") pod "openstack-operator-controller-manager-6c5c989645-kkkb7" (UID: "83f119ed-6576-4e47-82b1-a06189cff0a7") : secret "metrics-server-cert" not found Dec 03 11:09:02 crc kubenswrapper[4646]: E1203 11:09:02.036128 4646 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 11:09:02 crc kubenswrapper[4646]: E1203 11:09:02.036160 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs podName:83f119ed-6576-4e47-82b1-a06189cff0a7 nodeName:}" failed. No retries permitted until 2025-12-03 11:09:06.036153104 +0000 UTC m=+922.499209239 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs") pod "openstack-operator-controller-manager-6c5c989645-kkkb7" (UID: "83f119ed-6576-4e47-82b1-a06189cff0a7") : secret "webhook-server-cert" not found Dec 03 11:09:02 crc kubenswrapper[4646]: I1203 11:09:02.212444 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hblfs" event={"ID":"42a46b9b-c088-4e78-848e-c5b605694d3e","Type":"ContainerStarted","Data":"c39e46f00d4feee54213a251b18f08d7c18e7042bbdb59a8e1901deb2e656fde"} Dec 03 11:09:02 crc kubenswrapper[4646]: I1203 11:09:02.904329 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-f7pcq"] Dec 03 11:09:02 crc kubenswrapper[4646]: I1203 11:09:02.916375 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-l7v6k"] Dec 03 11:09:02 crc kubenswrapper[4646]: I1203 11:09:02.932209 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-w5c45"] Dec 03 11:09:03 crc kubenswrapper[4646]: I1203 11:09:03.247513 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-w5c45" event={"ID":"55e5ed61-839c-441b-acf2-cd43cc0c5c90","Type":"ContainerStarted","Data":"feccba7c98e7ef086e95aeee07e92456ac1972282d3a1d4254c18340b7f55b78"} Dec 03 11:09:03 crc kubenswrapper[4646]: I1203 11:09:03.254876 4646 generic.go:334] "Generic (PLEG): container finished" podID="b414a389-9e48-412c-b4e0-26a9b4cb51eb" containerID="e334477b13bc7ef7e1b3512a4a9b40bc77322c4da50bb327cf3228fcfc46cac4" exitCode=0 Dec 03 11:09:03 crc kubenswrapper[4646]: I1203 11:09:03.254953 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8pb7" event={"ID":"b414a389-9e48-412c-b4e0-26a9b4cb51eb","Type":"ContainerDied","Data":"e334477b13bc7ef7e1b3512a4a9b40bc77322c4da50bb327cf3228fcfc46cac4"} Dec 03 11:09:03 crc kubenswrapper[4646]: I1203 11:09:03.272975 4646 generic.go:334] "Generic (PLEG): container finished" podID="42a46b9b-c088-4e78-848e-c5b605694d3e" containerID="c559b9c19ff8f4aac03d7c72ac0736c992d01dcf395d05fc3d3fd9d978f225f4" exitCode=0 Dec 03 11:09:03 crc kubenswrapper[4646]: I1203 11:09:03.273031 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hblfs" event={"ID":"42a46b9b-c088-4e78-848e-c5b605694d3e","Type":"ContainerDied","Data":"c559b9c19ff8f4aac03d7c72ac0736c992d01dcf395d05fc3d3fd9d978f225f4"} Dec 03 11:09:03 crc kubenswrapper[4646]: I1203 11:09:03.279515 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-f7pcq" event={"ID":"011cd55c-5e76-4dd3-905b-62c587119096","Type":"ContainerStarted","Data":"435c5eeec6ed4f332cd5fdc43ceae07f0d29cf2d0b87fbf7679ebbafb6497597"} Dec 03 11:09:03 crc kubenswrapper[4646]: I1203 11:09:03.281290 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-l7v6k" event={"ID":"68e6047b-627f-4c45-aa55-4aeab90e17b5","Type":"ContainerStarted","Data":"ad49387931e9c117253754d0abab7304e42bd9458e3b83ec38dc8162bf2435c6"} Dec 03 11:09:03 crc kubenswrapper[4646]: I1203 11:09:03.295447 4646 generic.go:334] "Generic (PLEG): container finished" podID="682649bd-4003-4e88-a81a-dafee3703114" containerID="c28e07a54c869e2a8f2821107789d8cbf583f8113824777df445ba215819a6b3" exitCode=0 Dec 03 11:09:03 crc kubenswrapper[4646]: I1203 11:09:03.295475 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qspkq" event={"ID":"682649bd-4003-4e88-a81a-dafee3703114","Type":"ContainerDied","Data":"c28e07a54c869e2a8f2821107789d8cbf583f8113824777df445ba215819a6b3"} Dec 03 11:09:03 crc kubenswrapper[4646]: I1203 11:09:03.664963 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-rcmlq"] Dec 03 11:09:03 crc kubenswrapper[4646]: I1203 11:09:03.690686 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sc2vn"] Dec 03 11:09:03 crc kubenswrapper[4646]: I1203 11:09:03.711015 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-schbl"] Dec 03 11:09:03 crc kubenswrapper[4646]: W1203 11:09:03.742932 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3e07f77_c87f_4c6b_be4e_938d71354f8b.slice/crio-6e4ada006b427346c30861c3a82d307271d65189456159f605247b7849e1113c WatchSource:0}: Error finding container 6e4ada006b427346c30861c3a82d307271d65189456159f605247b7849e1113c: Status 404 returned error can't find the container with id 6e4ada006b427346c30861c3a82d307271d65189456159f605247b7849e1113c Dec 03 11:09:03 crc kubenswrapper[4646]: I1203 11:09:03.790990 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6hl22"] Dec 03 11:09:03 crc kubenswrapper[4646]: I1203 11:09:03.923227 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-hlwj7"] Dec 03 11:09:03 crc kubenswrapper[4646]: I1203 11:09:03.961741 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-t7fj8"] Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.005943 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-z6m86"] Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.024053 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mzlrq"] Dec 03 11:09:04 crc kubenswrapper[4646]: W1203 11:09:04.024514 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod396bd18c_37f7_471f_a0f9_0b9cb5a447e3.slice/crio-82b7280f859c9e3fff6518eaaf4c21695902a6a0bc9bc63d1d44d76337d88021 WatchSource:0}: Error finding container 82b7280f859c9e3fff6518eaaf4c21695902a6a0bc9bc63d1d44d76337d88021: Status 404 returned error can't find the container with id 82b7280f859c9e3fff6518eaaf4c21695902a6a0bc9bc63d1d44d76337d88021 Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.047097 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-92vgw"] Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.053673 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-j2bwb"] Dec 03 11:09:04 crc kubenswrapper[4646]: W1203 11:09:04.117548 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0fa06f6_534e_4ad4_806d_c13cad106809.slice/crio-91572bb336da4dbb7d58a95e87f0fd45bba16265f089ced9ebcbfd369979f4f5 WatchSource:0}: Error finding container 91572bb336da4dbb7d58a95e87f0fd45bba16265f089ced9ebcbfd369979f4f5: Status 404 returned error can't find the container with id 91572bb336da4dbb7d58a95e87f0fd45bba16265f089ced9ebcbfd369979f4f5 Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.146394 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8"] Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.213216 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8"] Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.230604 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm"] Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.258702 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5kmmc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-pnbk8_openstack-operators(d7ed2042-fe69-40f6-8440-49d63e851ffc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 11:09:04 crc kubenswrapper[4646]: W1203 11:09:04.265527 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fed03a1_2388_470c_843e_4de04fd6d9bc.slice/crio-7d485b4741b985c6cde0502aed97b0eba03aa15ba20ff92abc4af7611011aefa WatchSource:0}: Error finding container 7d485b4741b985c6cde0502aed97b0eba03aa15ba20ff92abc4af7611011aefa: Status 404 returned error can't find the container with id 7d485b4741b985c6cde0502aed97b0eba03aa15ba20ff92abc4af7611011aefa Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.265757 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tmpxh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-ss5tr_openstack-operators(e400aa20-0649-47de-8f10-da6090800eb3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.267469 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5vdh"] Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.272291 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dnmkm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-ps2zm_openstack-operators(4cb5b92a-540a-41d9-a662-b1f3faf07829): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.272681 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j4bk6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-g8pg8_openstack-operators(9a1416d1-243f-4ab9-9a9c-4b71e0f23876): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.283540 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5kmmc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-pnbk8_openstack-operators(d7ed2042-fe69-40f6-8440-49d63e851ffc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.284752 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" podUID="d7ed2042-fe69-40f6-8440-49d63e851ffc" Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.289106 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr"] Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.291902 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r764p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-56r4n_openstack-operators(1fed03a1-2388-470c-843e-4de04fd6d9bc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.294671 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tmpxh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-ss5tr_openstack-operators(e400aa20-0649-47de-8f10-da6090800eb3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.296009 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" podUID="e400aa20-0649-47de-8f10-da6090800eb3" Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.303383 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j4bk6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-g8pg8_openstack-operators(9a1416d1-243f-4ab9-9a9c-4b71e0f23876): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.303503 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r764p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-56r4n_openstack-operators(1fed03a1-2388-470c-843e-4de04fd6d9bc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.303591 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dnmkm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-ps2zm_openstack-operators(4cb5b92a-540a-41d9-a662-b1f3faf07829): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.304563 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" podUID="1fed03a1-2388-470c-843e-4de04fd6d9bc" Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.304621 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" podUID="9a1416d1-243f-4ab9-9a9c-4b71e0f23876" Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.304665 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" podUID="4cb5b92a-540a-41d9-a662-b1f3faf07829" Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.327806 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n"] Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.345268 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-n9s99"] Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.375282 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fs8bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-n9s99_openstack-operators(ed761af7-289e-4f6c-a7b9-710886250a3a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.377164 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fs8bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-n9s99_openstack-operators(ed761af7-289e-4f6c-a7b9-710886250a3a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.380329 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" podUID="ed761af7-289e-4f6c-a7b9-710886250a3a" Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.393658 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" event={"ID":"1fed03a1-2388-470c-843e-4de04fd6d9bc","Type":"ContainerStarted","Data":"7d485b4741b985c6cde0502aed97b0eba03aa15ba20ff92abc4af7611011aefa"} Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.395727 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hlwj7" event={"ID":"e2fc2280-6814-4ba0-bd48-2272d37f2b41","Type":"ContainerStarted","Data":"d1bb60c0a54a9e7730ac3e3100db45e43b0a89072045ac0e0b9e0dd1053de612"} Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.402692 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" podUID="1fed03a1-2388-470c-843e-4de04fd6d9bc" Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.405937 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-schbl" event={"ID":"31371552-226a-436f-a595-52bc387ed5cf","Type":"ContainerStarted","Data":"710dde39efe263f128954e3133dde18fcf9586778a88cdfcdb5885cccc3fd636"} Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.415729 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-t7fj8" event={"ID":"153f2bfa-6ca2-4a9f-900e-5321c5791639","Type":"ContainerStarted","Data":"1f7d254adfb3df334e1844622a2905706ce7616bfa6aee32f8f38fc6c27a01d4"} Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.417392 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" event={"ID":"9a1416d1-243f-4ab9-9a9c-4b71e0f23876","Type":"ContainerStarted","Data":"5b9c3e652289ed6b84fd5b8d7ddf263b75e62d1cec1392c1bf0e8f393504d624"} Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.425642 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" podUID="9a1416d1-243f-4ab9-9a9c-4b71e0f23876" Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.439965 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-j2bwb" event={"ID":"f274e7e0-a406-4617-95bc-fbca7ef39526","Type":"ContainerStarted","Data":"6556d34df2d1159e6602be63069608060771765a3409aa2a5efadb33d6615b06"} Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.467175 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5vdh" event={"ID":"afd2b4e9-59a4-4e52-bd72-706b8edbd719","Type":"ContainerStarted","Data":"267e02b03c3813b8ba43d42abb2bd221428ae1de0cf3910865a09949f3012c15"} Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.472316 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6hl22" event={"ID":"089f6f35-ab06-4b31-a331-16411a9783e7","Type":"ContainerStarted","Data":"67886862e693253e532cf0b7607095ab495b0eef55b9d8f3704f79fedddff065"} Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.491736 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" event={"ID":"ed761af7-289e-4f6c-a7b9-710886250a3a","Type":"ContainerStarted","Data":"6f2268d0fc3b4c242e25141584b1841e820ee58a6f807ceef18f2ef9ca86a62f"} Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.495626 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" podUID="ed761af7-289e-4f6c-a7b9-710886250a3a" Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.511914 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-92vgw" event={"ID":"a0fa06f6-534e-4ad4-806d-c13cad106809","Type":"ContainerStarted","Data":"91572bb336da4dbb7d58a95e87f0fd45bba16265f089ced9ebcbfd369979f4f5"} Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.518478 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6m86" event={"ID":"0be433f8-aae1-48a4-8853-de55ae48f836","Type":"ContainerStarted","Data":"e2c5595ba36d94e1e49221f6ea121cc86301e85a0cfc8d50368fe04a30a619b1"} Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.525283 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sc2vn" event={"ID":"eaefadf5-75cd-4ff2-811c-e25bb2850173","Type":"ContainerStarted","Data":"57774748a9ff24950bb60fabe24833c63880e61c312e4604122c103ee7704e1f"} Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.530885 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mzlrq" event={"ID":"396bd18c-37f7-471f-a0f9-0b9cb5a447e3","Type":"ContainerStarted","Data":"82b7280f859c9e3fff6518eaaf4c21695902a6a0bc9bc63d1d44d76337d88021"} Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.538769 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rcmlq" event={"ID":"f3e07f77-c87f-4c6b-be4e-938d71354f8b","Type":"ContainerStarted","Data":"6e4ada006b427346c30861c3a82d307271d65189456159f605247b7849e1113c"} Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.547096 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" event={"ID":"d7ed2042-fe69-40f6-8440-49d63e851ffc","Type":"ContainerStarted","Data":"1ed230f83c0859b78a74ab1dc477f60ca0e08bd0332c4786946a06e7b65cfab2"} Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.552653 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" podUID="d7ed2042-fe69-40f6-8440-49d63e851ffc" Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.554530 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" event={"ID":"4cb5b92a-540a-41d9-a662-b1f3faf07829","Type":"ContainerStarted","Data":"8b2db2f5b81d60f8c74de42541acdb72df7c95126a8d996c9596606ae3101efc"} Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.569583 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" podUID="4cb5b92a-540a-41d9-a662-b1f3faf07829" Dec 03 11:09:04 crc kubenswrapper[4646]: I1203 11:09:04.574004 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" event={"ID":"e400aa20-0649-47de-8f10-da6090800eb3","Type":"ContainerStarted","Data":"4bcde65bc15ea1ed525271ad9c835b76fdc4f67b11fe9445b771245353b57043"} Dec 03 11:09:04 crc kubenswrapper[4646]: E1203 11:09:04.578138 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" podUID="e400aa20-0649-47de-8f10-da6090800eb3" Dec 03 11:09:05 crc kubenswrapper[4646]: I1203 11:09:05.429057 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert\") pod \"infra-operator-controller-manager-57548d458d-g6jbp\" (UID: \"d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" Dec 03 11:09:05 crc kubenswrapper[4646]: E1203 11:09:05.429325 4646 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 11:09:05 crc kubenswrapper[4646]: E1203 11:09:05.429583 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert podName:d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0 nodeName:}" failed. No retries permitted until 2025-12-03 11:09:13.429560936 +0000 UTC m=+929.892617161 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert") pod "infra-operator-controller-manager-57548d458d-g6jbp" (UID: "d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0") : secret "infra-operator-webhook-server-cert" not found Dec 03 11:09:05 crc kubenswrapper[4646]: I1203 11:09:05.614159 4646 generic.go:334] "Generic (PLEG): container finished" podID="42a46b9b-c088-4e78-848e-c5b605694d3e" containerID="b36fc4103cec0f9bdd67113e9eda575446560e31f8951000d1f93e683bbacab6" exitCode=0 Dec 03 11:09:05 crc kubenswrapper[4646]: I1203 11:09:05.614230 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hblfs" event={"ID":"42a46b9b-c088-4e78-848e-c5b605694d3e","Type":"ContainerDied","Data":"b36fc4103cec0f9bdd67113e9eda575446560e31f8951000d1f93e683bbacab6"} Dec 03 11:09:05 crc kubenswrapper[4646]: I1203 11:09:05.621067 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qspkq" event={"ID":"682649bd-4003-4e88-a81a-dafee3703114","Type":"ContainerStarted","Data":"703c3b963d425f0fa77c7cc9ac30ad216ee76cfa778f987cad6af7be62c63de7"} Dec 03 11:09:05 crc kubenswrapper[4646]: I1203 11:09:05.644860 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8pb7" event={"ID":"b414a389-9e48-412c-b4e0-26a9b4cb51eb","Type":"ContainerStarted","Data":"430d69cae0f9e978077e5affa01ee4bc0866f1712455d182c63f9b00c123e2c3"} Dec 03 11:09:05 crc kubenswrapper[4646]: I1203 11:09:05.664968 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qspkq" podStartSLOduration=4.257286819 podStartE2EDuration="20.664949682s" podCreationTimestamp="2025-12-03 11:08:45 +0000 UTC" firstStartedPulling="2025-12-03 11:08:47.968721973 +0000 UTC m=+904.431778108" lastFinishedPulling="2025-12-03 11:09:04.376384836 +0000 UTC m=+920.839440971" observedRunningTime="2025-12-03 11:09:05.662129961 +0000 UTC m=+922.125186106" watchObservedRunningTime="2025-12-03 11:09:05.664949682 +0000 UTC m=+922.128005817" Dec 03 11:09:05 crc kubenswrapper[4646]: E1203 11:09:05.677710 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" podUID="1fed03a1-2388-470c-843e-4de04fd6d9bc" Dec 03 11:09:05 crc kubenswrapper[4646]: E1203 11:09:05.677841 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" podUID="4cb5b92a-540a-41d9-a662-b1f3faf07829" Dec 03 11:09:05 crc kubenswrapper[4646]: E1203 11:09:05.677903 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" podUID="ed761af7-289e-4f6c-a7b9-710886250a3a" Dec 03 11:09:05 crc kubenswrapper[4646]: E1203 11:09:05.677962 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" podUID="e400aa20-0649-47de-8f10-da6090800eb3" Dec 03 11:09:05 crc kubenswrapper[4646]: E1203 11:09:05.682788 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" podUID="9a1416d1-243f-4ab9-9a9c-4b71e0f23876" Dec 03 11:09:05 crc kubenswrapper[4646]: E1203 11:09:05.684728 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" podUID="d7ed2042-fe69-40f6-8440-49d63e851ffc" Dec 03 11:09:05 crc kubenswrapper[4646]: I1203 11:09:05.753621 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9\" (UID: \"e776d6da-584d-498f-b611-0eb67be6d6d0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" Dec 03 11:09:05 crc kubenswrapper[4646]: E1203 11:09:05.758065 4646 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 11:09:05 crc kubenswrapper[4646]: E1203 11:09:05.758133 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert podName:e776d6da-584d-498f-b611-0eb67be6d6d0 nodeName:}" failed. No retries permitted until 2025-12-03 11:09:13.758114026 +0000 UTC m=+930.221170231 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" (UID: "e776d6da-584d-498f-b611-0eb67be6d6d0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 11:09:06 crc kubenswrapper[4646]: I1203 11:09:06.056918 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:09:06 crc kubenswrapper[4646]: I1203 11:09:06.057361 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:09:06 crc kubenswrapper[4646]: E1203 11:09:06.057423 4646 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 11:09:06 crc kubenswrapper[4646]: E1203 11:09:06.057491 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs podName:83f119ed-6576-4e47-82b1-a06189cff0a7 nodeName:}" failed. No retries permitted until 2025-12-03 11:09:14.057473271 +0000 UTC m=+930.520529396 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs") pod "openstack-operator-controller-manager-6c5c989645-kkkb7" (UID: "83f119ed-6576-4e47-82b1-a06189cff0a7") : secret "webhook-server-cert" not found Dec 03 11:09:06 crc kubenswrapper[4646]: E1203 11:09:06.057495 4646 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 11:09:06 crc kubenswrapper[4646]: E1203 11:09:06.057543 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs podName:83f119ed-6576-4e47-82b1-a06189cff0a7 nodeName:}" failed. No retries permitted until 2025-12-03 11:09:14.057528263 +0000 UTC m=+930.520584398 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs") pod "openstack-operator-controller-manager-6c5c989645-kkkb7" (UID: "83f119ed-6576-4e47-82b1-a06189cff0a7") : secret "metrics-server-cert" not found Dec 03 11:09:06 crc kubenswrapper[4646]: I1203 11:09:06.677045 4646 generic.go:334] "Generic (PLEG): container finished" podID="b414a389-9e48-412c-b4e0-26a9b4cb51eb" containerID="430d69cae0f9e978077e5affa01ee4bc0866f1712455d182c63f9b00c123e2c3" exitCode=0 Dec 03 11:09:06 crc kubenswrapper[4646]: I1203 11:09:06.678153 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8pb7" event={"ID":"b414a389-9e48-412c-b4e0-26a9b4cb51eb","Type":"ContainerDied","Data":"430d69cae0f9e978077e5affa01ee4bc0866f1712455d182c63f9b00c123e2c3"} Dec 03 11:09:08 crc kubenswrapper[4646]: I1203 11:09:08.715783 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hblfs" event={"ID":"42a46b9b-c088-4e78-848e-c5b605694d3e","Type":"ContainerStarted","Data":"0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0"} Dec 03 11:09:08 crc kubenswrapper[4646]: I1203 11:09:08.735739 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hblfs" podStartSLOduration=11.593521835 podStartE2EDuration="15.735721896s" podCreationTimestamp="2025-12-03 11:08:53 +0000 UTC" firstStartedPulling="2025-12-03 11:09:03.276436085 +0000 UTC m=+919.739492220" lastFinishedPulling="2025-12-03 11:09:07.418636146 +0000 UTC m=+923.881692281" observedRunningTime="2025-12-03 11:09:08.734777399 +0000 UTC m=+925.197833534" watchObservedRunningTime="2025-12-03 11:09:08.735721896 +0000 UTC m=+925.198778031" Dec 03 11:09:10 crc kubenswrapper[4646]: I1203 11:09:10.737039 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8pb7" event={"ID":"b414a389-9e48-412c-b4e0-26a9b4cb51eb","Type":"ContainerStarted","Data":"4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676"} Dec 03 11:09:10 crc kubenswrapper[4646]: I1203 11:09:10.790478 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n8pb7" podStartSLOduration=16.166252197 podStartE2EDuration="20.790463954s" podCreationTimestamp="2025-12-03 11:08:50 +0000 UTC" firstStartedPulling="2025-12-03 11:09:03.257908329 +0000 UTC m=+919.720964464" lastFinishedPulling="2025-12-03 11:09:07.882120086 +0000 UTC m=+924.345176221" observedRunningTime="2025-12-03 11:09:10.788270501 +0000 UTC m=+927.251326636" watchObservedRunningTime="2025-12-03 11:09:10.790463954 +0000 UTC m=+927.253520089" Dec 03 11:09:11 crc kubenswrapper[4646]: I1203 11:09:11.361937 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:09:11 crc kubenswrapper[4646]: I1203 11:09:11.361978 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:09:12 crc kubenswrapper[4646]: I1203 11:09:12.401460 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-n8pb7" podUID="b414a389-9e48-412c-b4e0-26a9b4cb51eb" containerName="registry-server" probeResult="failure" output=< Dec 03 11:09:12 crc kubenswrapper[4646]: timeout: failed to connect service ":50051" within 1s Dec 03 11:09:12 crc kubenswrapper[4646]: > Dec 03 11:09:13 crc kubenswrapper[4646]: I1203 11:09:13.463499 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert\") pod \"infra-operator-controller-manager-57548d458d-g6jbp\" (UID: \"d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" Dec 03 11:09:13 crc kubenswrapper[4646]: I1203 11:09:13.470112 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0-cert\") pod \"infra-operator-controller-manager-57548d458d-g6jbp\" (UID: \"d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" Dec 03 11:09:13 crc kubenswrapper[4646]: I1203 11:09:13.556988 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" Dec 03 11:09:13 crc kubenswrapper[4646]: I1203 11:09:13.768038 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9\" (UID: \"e776d6da-584d-498f-b611-0eb67be6d6d0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" Dec 03 11:09:13 crc kubenswrapper[4646]: I1203 11:09:13.772085 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e776d6da-584d-498f-b611-0eb67be6d6d0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9\" (UID: \"e776d6da-584d-498f-b611-0eb67be6d6d0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" Dec 03 11:09:13 crc kubenswrapper[4646]: I1203 11:09:13.908262 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" Dec 03 11:09:14 crc kubenswrapper[4646]: I1203 11:09:14.072229 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:09:14 crc kubenswrapper[4646]: I1203 11:09:14.072288 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:09:14 crc kubenswrapper[4646]: E1203 11:09:14.072426 4646 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 11:09:14 crc kubenswrapper[4646]: E1203 11:09:14.072483 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs podName:83f119ed-6576-4e47-82b1-a06189cff0a7 nodeName:}" failed. No retries permitted until 2025-12-03 11:09:30.072468454 +0000 UTC m=+946.535524589 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs") pod "openstack-operator-controller-manager-6c5c989645-kkkb7" (UID: "83f119ed-6576-4e47-82b1-a06189cff0a7") : secret "webhook-server-cert" not found Dec 03 11:09:14 crc kubenswrapper[4646]: I1203 11:09:14.075026 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-metrics-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:09:14 crc kubenswrapper[4646]: I1203 11:09:14.207159 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:09:14 crc kubenswrapper[4646]: I1203 11:09:14.207212 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:09:14 crc kubenswrapper[4646]: I1203 11:09:14.251071 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:09:14 crc kubenswrapper[4646]: I1203 11:09:14.821923 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:09:14 crc kubenswrapper[4646]: I1203 11:09:14.865759 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hblfs"] Dec 03 11:09:15 crc kubenswrapper[4646]: I1203 11:09:15.466579 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qspkq" Dec 03 11:09:15 crc kubenswrapper[4646]: I1203 11:09:15.466618 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qspkq" Dec 03 11:09:15 crc kubenswrapper[4646]: I1203 11:09:15.547466 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qspkq" Dec 03 11:09:15 crc kubenswrapper[4646]: I1203 11:09:15.803716 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qspkq" Dec 03 11:09:16 crc kubenswrapper[4646]: I1203 11:09:16.720737 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qspkq"] Dec 03 11:09:16 crc kubenswrapper[4646]: I1203 11:09:16.773309 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hblfs" podUID="42a46b9b-c088-4e78-848e-c5b605694d3e" containerName="registry-server" containerID="cri-o://0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0" gracePeriod=2 Dec 03 11:09:16 crc kubenswrapper[4646]: I1203 11:09:16.884280 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rz4r5"] Dec 03 11:09:16 crc kubenswrapper[4646]: I1203 11:09:16.884534 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rz4r5" podUID="75816a6e-5fea-4212-964b-aafb2dce1385" containerName="registry-server" containerID="cri-o://9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b" gracePeriod=2 Dec 03 11:09:17 crc kubenswrapper[4646]: I1203 11:09:17.781364 4646 generic.go:334] "Generic (PLEG): container finished" podID="42a46b9b-c088-4e78-848e-c5b605694d3e" containerID="0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0" exitCode=0 Dec 03 11:09:17 crc kubenswrapper[4646]: I1203 11:09:17.781435 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hblfs" event={"ID":"42a46b9b-c088-4e78-848e-c5b605694d3e","Type":"ContainerDied","Data":"0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0"} Dec 03 11:09:17 crc kubenswrapper[4646]: I1203 11:09:17.784198 4646 generic.go:334] "Generic (PLEG): container finished" podID="75816a6e-5fea-4212-964b-aafb2dce1385" containerID="9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b" exitCode=0 Dec 03 11:09:17 crc kubenswrapper[4646]: I1203 11:09:17.784252 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rz4r5" event={"ID":"75816a6e-5fea-4212-964b-aafb2dce1385","Type":"ContainerDied","Data":"9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b"} Dec 03 11:09:18 crc kubenswrapper[4646]: E1203 11:09:18.698809 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 03 11:09:18 crc kubenswrapper[4646]: E1203 11:09:18.698999 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9qjsn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-sc2vn_openstack-operators(eaefadf5-75cd-4ff2-811c-e25bb2850173): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:20 crc kubenswrapper[4646]: E1203 11:09:20.252179 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 03 11:09:20 crc kubenswrapper[4646]: E1203 11:09:20.252812 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4wd4d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-92vgw_openstack-operators(a0fa06f6-534e-4ad4-806d-c13cad106809): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:21 crc kubenswrapper[4646]: I1203 11:09:21.423981 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:09:21 crc kubenswrapper[4646]: I1203 11:09:21.502586 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:09:22 crc kubenswrapper[4646]: I1203 11:09:22.218506 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n8pb7"] Dec 03 11:09:23 crc kubenswrapper[4646]: I1203 11:09:23.146541 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n8pb7" podUID="b414a389-9e48-412c-b4e0-26a9b4cb51eb" containerName="registry-server" containerID="cri-o://4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676" gracePeriod=2 Dec 03 11:09:23 crc kubenswrapper[4646]: E1203 11:09:23.407312 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b is running failed: container process not found" containerID="9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:23 crc kubenswrapper[4646]: E1203 11:09:23.408170 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b is running failed: container process not found" containerID="9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:23 crc kubenswrapper[4646]: E1203 11:09:23.408944 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b is running failed: container process not found" containerID="9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:23 crc kubenswrapper[4646]: E1203 11:09:23.408973 4646 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-rz4r5" podUID="75816a6e-5fea-4212-964b-aafb2dce1385" containerName="registry-server" Dec 03 11:09:23 crc kubenswrapper[4646]: E1203 11:09:23.874965 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385" Dec 03 11:09:23 crc kubenswrapper[4646]: E1203 11:09:23.875154 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-h6g6n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-w5c45_openstack-operators(55e5ed61-839c-441b-acf2-cd43cc0c5c90): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:24 crc kubenswrapper[4646]: I1203 11:09:24.157175 4646 generic.go:334] "Generic (PLEG): container finished" podID="b414a389-9e48-412c-b4e0-26a9b4cb51eb" containerID="4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676" exitCode=0 Dec 03 11:09:24 crc kubenswrapper[4646]: I1203 11:09:24.157218 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8pb7" event={"ID":"b414a389-9e48-412c-b4e0-26a9b4cb51eb","Type":"ContainerDied","Data":"4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676"} Dec 03 11:09:24 crc kubenswrapper[4646]: E1203 11:09:24.208215 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0 is running failed: container process not found" containerID="0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:24 crc kubenswrapper[4646]: E1203 11:09:24.208998 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0 is running failed: container process not found" containerID="0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:24 crc kubenswrapper[4646]: E1203 11:09:24.209393 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0 is running failed: container process not found" containerID="0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:24 crc kubenswrapper[4646]: E1203 11:09:24.209440 4646 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-hblfs" podUID="42a46b9b-c088-4e78-848e-c5b605694d3e" containerName="registry-server" Dec 03 11:09:26 crc kubenswrapper[4646]: E1203 11:09:26.371936 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 03 11:09:26 crc kubenswrapper[4646]: E1203 11:09:26.372397 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mt528,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-f7pcq_openstack-operators(011cd55c-5e76-4dd3-905b-62c587119096): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:26 crc kubenswrapper[4646]: I1203 11:09:26.376976 4646 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 11:09:27 crc kubenswrapper[4646]: E1203 11:09:27.145106 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 03 11:09:27 crc kubenswrapper[4646]: E1203 11:09:27.145317 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p4qxc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-6hl22_openstack-operators(089f6f35-ab06-4b31-a331-16411a9783e7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:29 crc kubenswrapper[4646]: E1203 11:09:29.101440 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 03 11:09:29 crc kubenswrapper[4646]: E1203 11:09:29.102700 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2nsls,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-j2bwb_openstack-operators(f274e7e0-a406-4617-95bc-fbca7ef39526): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:30 crc kubenswrapper[4646]: I1203 11:09:30.167729 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:09:30 crc kubenswrapper[4646]: I1203 11:09:30.175684 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/83f119ed-6576-4e47-82b1-a06189cff0a7-webhook-certs\") pod \"openstack-operator-controller-manager-6c5c989645-kkkb7\" (UID: \"83f119ed-6576-4e47-82b1-a06189cff0a7\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:09:30 crc kubenswrapper[4646]: I1203 11:09:30.476767 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:09:30 crc kubenswrapper[4646]: E1203 11:09:30.825259 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 03 11:09:30 crc kubenswrapper[4646]: E1203 11:09:30.825437 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-47k8t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-t7fj8_openstack-operators(153f2bfa-6ca2-4a9f-900e-5321c5791639): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:31 crc kubenswrapper[4646]: E1203 11:09:31.362502 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676 is running failed: container process not found" containerID="4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:31 crc kubenswrapper[4646]: E1203 11:09:31.362797 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676 is running failed: container process not found" containerID="4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:31 crc kubenswrapper[4646]: E1203 11:09:31.363018 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676 is running failed: container process not found" containerID="4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:31 crc kubenswrapper[4646]: E1203 11:09:31.363044 4646 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-n8pb7" podUID="b414a389-9e48-412c-b4e0-26a9b4cb51eb" containerName="registry-server" Dec 03 11:09:31 crc kubenswrapper[4646]: E1203 11:09:31.891192 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 03 11:09:31 crc kubenswrapper[4646]: E1203 11:09:31.891431 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-npqn5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-schbl_openstack-operators(31371552-226a-436f-a595-52bc387ed5cf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:32 crc kubenswrapper[4646]: E1203 11:09:32.644783 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 03 11:09:32 crc kubenswrapper[4646]: E1203 11:09:32.645288 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-st64w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-mzlrq_openstack-operators(396bd18c-37f7-471f-a0f9-0b9cb5a447e3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:33 crc kubenswrapper[4646]: E1203 11:09:33.406173 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b is running failed: container process not found" containerID="9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:33 crc kubenswrapper[4646]: E1203 11:09:33.406937 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b is running failed: container process not found" containerID="9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:33 crc kubenswrapper[4646]: E1203 11:09:33.407254 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b is running failed: container process not found" containerID="9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:33 crc kubenswrapper[4646]: E1203 11:09:33.407301 4646 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-rz4r5" podUID="75816a6e-5fea-4212-964b-aafb2dce1385" containerName="registry-server" Dec 03 11:09:34 crc kubenswrapper[4646]: E1203 11:09:34.207569 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0 is running failed: container process not found" containerID="0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:34 crc kubenswrapper[4646]: E1203 11:09:34.208092 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0 is running failed: container process not found" containerID="0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:34 crc kubenswrapper[4646]: E1203 11:09:34.208546 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0 is running failed: container process not found" containerID="0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:34 crc kubenswrapper[4646]: E1203 11:09:34.208583 4646 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-hblfs" podUID="42a46b9b-c088-4e78-848e-c5b605694d3e" containerName="registry-server" Dec 03 11:09:34 crc kubenswrapper[4646]: E1203 11:09:34.222866 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9" Dec 03 11:09:34 crc kubenswrapper[4646]: E1203 11:09:34.223393 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wmnwd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-rcmlq_openstack-operators(f3e07f77-c87f-4c6b-be4e-938d71354f8b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:35 crc kubenswrapper[4646]: E1203 11:09:35.266709 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 03 11:09:35 crc kubenswrapper[4646]: E1203 11:09:35.266898 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tr459,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-z6m86_openstack-operators(0be433f8-aae1-48a4-8853-de55ae48f836): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:36 crc kubenswrapper[4646]: I1203 11:09:36.448283 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:09:36 crc kubenswrapper[4646]: I1203 11:09:36.481082 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgh6r\" (UniqueName: \"kubernetes.io/projected/42a46b9b-c088-4e78-848e-c5b605694d3e-kube-api-access-lgh6r\") pod \"42a46b9b-c088-4e78-848e-c5b605694d3e\" (UID: \"42a46b9b-c088-4e78-848e-c5b605694d3e\") " Dec 03 11:09:36 crc kubenswrapper[4646]: I1203 11:09:36.481273 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42a46b9b-c088-4e78-848e-c5b605694d3e-utilities\") pod \"42a46b9b-c088-4e78-848e-c5b605694d3e\" (UID: \"42a46b9b-c088-4e78-848e-c5b605694d3e\") " Dec 03 11:09:36 crc kubenswrapper[4646]: I1203 11:09:36.481315 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42a46b9b-c088-4e78-848e-c5b605694d3e-catalog-content\") pod \"42a46b9b-c088-4e78-848e-c5b605694d3e\" (UID: \"42a46b9b-c088-4e78-848e-c5b605694d3e\") " Dec 03 11:09:36 crc kubenswrapper[4646]: I1203 11:09:36.488663 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42a46b9b-c088-4e78-848e-c5b605694d3e-utilities" (OuterVolumeSpecName: "utilities") pod "42a46b9b-c088-4e78-848e-c5b605694d3e" (UID: "42a46b9b-c088-4e78-848e-c5b605694d3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:09:36 crc kubenswrapper[4646]: I1203 11:09:36.504565 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42a46b9b-c088-4e78-848e-c5b605694d3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42a46b9b-c088-4e78-848e-c5b605694d3e" (UID: "42a46b9b-c088-4e78-848e-c5b605694d3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:09:36 crc kubenswrapper[4646]: I1203 11:09:36.507367 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42a46b9b-c088-4e78-848e-c5b605694d3e-kube-api-access-lgh6r" (OuterVolumeSpecName: "kube-api-access-lgh6r") pod "42a46b9b-c088-4e78-848e-c5b605694d3e" (UID: "42a46b9b-c088-4e78-848e-c5b605694d3e"). InnerVolumeSpecName "kube-api-access-lgh6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:09:36 crc kubenswrapper[4646]: I1203 11:09:36.583302 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgh6r\" (UniqueName: \"kubernetes.io/projected/42a46b9b-c088-4e78-848e-c5b605694d3e-kube-api-access-lgh6r\") on node \"crc\" DevicePath \"\"" Dec 03 11:09:36 crc kubenswrapper[4646]: I1203 11:09:36.583354 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42a46b9b-c088-4e78-848e-c5b605694d3e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 11:09:36 crc kubenswrapper[4646]: I1203 11:09:36.583364 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42a46b9b-c088-4e78-848e-c5b605694d3e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 11:09:36 crc kubenswrapper[4646]: I1203 11:09:36.928388 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9"] Dec 03 11:09:37 crc kubenswrapper[4646]: E1203 11:09:37.159324 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 03 11:09:37 crc kubenswrapper[4646]: E1203 11:09:37.159535 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cfvsz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-b5vdh_openstack-operators(afd2b4e9-59a4-4e52-bd72-706b8edbd719): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:37 crc kubenswrapper[4646]: E1203 11:09:37.161445 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5vdh" podUID="afd2b4e9-59a4-4e52-bd72-706b8edbd719" Dec 03 11:09:37 crc kubenswrapper[4646]: I1203 11:09:37.273596 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hblfs" event={"ID":"42a46b9b-c088-4e78-848e-c5b605694d3e","Type":"ContainerDied","Data":"c39e46f00d4feee54213a251b18f08d7c18e7042bbdb59a8e1901deb2e656fde"} Dec 03 11:09:37 crc kubenswrapper[4646]: I1203 11:09:37.273648 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hblfs" Dec 03 11:09:37 crc kubenswrapper[4646]: I1203 11:09:37.273661 4646 scope.go:117] "RemoveContainer" containerID="0433275da3c1897bf7ca2a6d6ceee77e1b2c4cc767cbba98207785c485e472e0" Dec 03 11:09:37 crc kubenswrapper[4646]: E1203 11:09:37.275898 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5vdh" podUID="afd2b4e9-59a4-4e52-bd72-706b8edbd719" Dec 03 11:09:37 crc kubenswrapper[4646]: I1203 11:09:37.307777 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hblfs"] Dec 03 11:09:37 crc kubenswrapper[4646]: I1203 11:09:37.312177 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hblfs"] Dec 03 11:09:37 crc kubenswrapper[4646]: E1203 11:09:37.785456 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801" Dec 03 11:09:37 crc kubenswrapper[4646]: E1203 11:09:37.785648 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5kmmc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-pnbk8_openstack-operators(d7ed2042-fe69-40f6-8440-49d63e851ffc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:37 crc kubenswrapper[4646]: I1203 11:09:37.867538 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42a46b9b-c088-4e78-848e-c5b605694d3e" path="/var/lib/kubelet/pods/42a46b9b-c088-4e78-848e-c5b605694d3e/volumes" Dec 03 11:09:41 crc kubenswrapper[4646]: E1203 11:09:41.363401 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676 is running failed: container process not found" containerID="4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:41 crc kubenswrapper[4646]: E1203 11:09:41.363929 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676 is running failed: container process not found" containerID="4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:41 crc kubenswrapper[4646]: E1203 11:09:41.364236 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676 is running failed: container process not found" containerID="4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:41 crc kubenswrapper[4646]: E1203 11:09:41.364264 4646 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-n8pb7" podUID="b414a389-9e48-412c-b4e0-26a9b4cb51eb" containerName="registry-server" Dec 03 11:09:41 crc kubenswrapper[4646]: E1203 11:09:41.622677 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 03 11:09:41 crc kubenswrapper[4646]: E1203 11:09:41.622891 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fs8bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-n9s99_openstack-operators(ed761af7-289e-4f6c-a7b9-710886250a3a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:42 crc kubenswrapper[4646]: E1203 11:09:42.781646 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809" Dec 03 11:09:42 crc kubenswrapper[4646]: E1203 11:09:42.782388 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r764p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-56r4n_openstack-operators(1fed03a1-2388-470c-843e-4de04fd6d9bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:43 crc kubenswrapper[4646]: E1203 11:09:43.405523 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b is running failed: container process not found" containerID="9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:43 crc kubenswrapper[4646]: E1203 11:09:43.405966 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b is running failed: container process not found" containerID="9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:43 crc kubenswrapper[4646]: E1203 11:09:43.406162 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b is running failed: container process not found" containerID="9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 11:09:43 crc kubenswrapper[4646]: E1203 11:09:43.406234 4646 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-rz4r5" podUID="75816a6e-5fea-4212-964b-aafb2dce1385" containerName="registry-server" Dec 03 11:09:43 crc kubenswrapper[4646]: E1203 11:09:43.892383 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 03 11:09:43 crc kubenswrapper[4646]: E1203 11:09:43.893428 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tmpxh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-ss5tr_openstack-operators(e400aa20-0649-47de-8f10-da6090800eb3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:45 crc kubenswrapper[4646]: E1203 11:09:45.538716 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85" Dec 03 11:09:45 crc kubenswrapper[4646]: E1203 11:09:45.538975 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dnmkm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-ps2zm_openstack-operators(4cb5b92a-540a-41d9-a662-b1f3faf07829): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:45 crc kubenswrapper[4646]: I1203 11:09:45.597305 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 11:09:45 crc kubenswrapper[4646]: I1203 11:09:45.714088 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rk4fq\" (UniqueName: \"kubernetes.io/projected/75816a6e-5fea-4212-964b-aafb2dce1385-kube-api-access-rk4fq\") pod \"75816a6e-5fea-4212-964b-aafb2dce1385\" (UID: \"75816a6e-5fea-4212-964b-aafb2dce1385\") " Dec 03 11:09:45 crc kubenswrapper[4646]: I1203 11:09:45.714159 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75816a6e-5fea-4212-964b-aafb2dce1385-catalog-content\") pod \"75816a6e-5fea-4212-964b-aafb2dce1385\" (UID: \"75816a6e-5fea-4212-964b-aafb2dce1385\") " Dec 03 11:09:45 crc kubenswrapper[4646]: I1203 11:09:45.714369 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75816a6e-5fea-4212-964b-aafb2dce1385-utilities\") pod \"75816a6e-5fea-4212-964b-aafb2dce1385\" (UID: \"75816a6e-5fea-4212-964b-aafb2dce1385\") " Dec 03 11:09:45 crc kubenswrapper[4646]: I1203 11:09:45.717706 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75816a6e-5fea-4212-964b-aafb2dce1385-utilities" (OuterVolumeSpecName: "utilities") pod "75816a6e-5fea-4212-964b-aafb2dce1385" (UID: "75816a6e-5fea-4212-964b-aafb2dce1385"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:09:45 crc kubenswrapper[4646]: I1203 11:09:45.737692 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75816a6e-5fea-4212-964b-aafb2dce1385-kube-api-access-rk4fq" (OuterVolumeSpecName: "kube-api-access-rk4fq") pod "75816a6e-5fea-4212-964b-aafb2dce1385" (UID: "75816a6e-5fea-4212-964b-aafb2dce1385"). InnerVolumeSpecName "kube-api-access-rk4fq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:09:45 crc kubenswrapper[4646]: I1203 11:09:45.768866 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75816a6e-5fea-4212-964b-aafb2dce1385-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75816a6e-5fea-4212-964b-aafb2dce1385" (UID: "75816a6e-5fea-4212-964b-aafb2dce1385"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:09:45 crc kubenswrapper[4646]: I1203 11:09:45.816553 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rk4fq\" (UniqueName: \"kubernetes.io/projected/75816a6e-5fea-4212-964b-aafb2dce1385-kube-api-access-rk4fq\") on node \"crc\" DevicePath \"\"" Dec 03 11:09:45 crc kubenswrapper[4646]: I1203 11:09:45.816584 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75816a6e-5fea-4212-964b-aafb2dce1385-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 11:09:45 crc kubenswrapper[4646]: I1203 11:09:45.816600 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75816a6e-5fea-4212-964b-aafb2dce1385-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 11:09:46 crc kubenswrapper[4646]: I1203 11:09:46.337327 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rz4r5" event={"ID":"75816a6e-5fea-4212-964b-aafb2dce1385","Type":"ContainerDied","Data":"dc5b0bd19899617f840dc714cbefc1ed9700627375a2e3360f65313030f65cdc"} Dec 03 11:09:46 crc kubenswrapper[4646]: I1203 11:09:46.337535 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rz4r5" Dec 03 11:09:46 crc kubenswrapper[4646]: I1203 11:09:46.359983 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rz4r5"] Dec 03 11:09:46 crc kubenswrapper[4646]: I1203 11:09:46.368530 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rz4r5"] Dec 03 11:09:46 crc kubenswrapper[4646]: E1203 11:09:46.377251 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 03 11:09:46 crc kubenswrapper[4646]: E1203 11:09:46.377673 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gnf4m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-l7v6k_openstack-operators(68e6047b-627f-4c45-aa55-4aeab90e17b5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:46 crc kubenswrapper[4646]: I1203 11:09:46.398150 4646 scope.go:117] "RemoveContainer" containerID="b36fc4103cec0f9bdd67113e9eda575446560e31f8951000d1f93e683bbacab6" Dec 03 11:09:46 crc kubenswrapper[4646]: I1203 11:09:46.432304 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:09:46 crc kubenswrapper[4646]: I1203 11:09:46.627092 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b414a389-9e48-412c-b4e0-26a9b4cb51eb-catalog-content\") pod \"b414a389-9e48-412c-b4e0-26a9b4cb51eb\" (UID: \"b414a389-9e48-412c-b4e0-26a9b4cb51eb\") " Dec 03 11:09:46 crc kubenswrapper[4646]: I1203 11:09:46.627168 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b414a389-9e48-412c-b4e0-26a9b4cb51eb-utilities\") pod \"b414a389-9e48-412c-b4e0-26a9b4cb51eb\" (UID: \"b414a389-9e48-412c-b4e0-26a9b4cb51eb\") " Dec 03 11:09:46 crc kubenswrapper[4646]: I1203 11:09:46.627292 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjz45\" (UniqueName: \"kubernetes.io/projected/b414a389-9e48-412c-b4e0-26a9b4cb51eb-kube-api-access-hjz45\") pod \"b414a389-9e48-412c-b4e0-26a9b4cb51eb\" (UID: \"b414a389-9e48-412c-b4e0-26a9b4cb51eb\") " Dec 03 11:09:46 crc kubenswrapper[4646]: I1203 11:09:46.627907 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b414a389-9e48-412c-b4e0-26a9b4cb51eb-utilities" (OuterVolumeSpecName: "utilities") pod "b414a389-9e48-412c-b4e0-26a9b4cb51eb" (UID: "b414a389-9e48-412c-b4e0-26a9b4cb51eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:09:46 crc kubenswrapper[4646]: I1203 11:09:46.631402 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b414a389-9e48-412c-b4e0-26a9b4cb51eb-kube-api-access-hjz45" (OuterVolumeSpecName: "kube-api-access-hjz45") pod "b414a389-9e48-412c-b4e0-26a9b4cb51eb" (UID: "b414a389-9e48-412c-b4e0-26a9b4cb51eb"). InnerVolumeSpecName "kube-api-access-hjz45". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:09:46 crc kubenswrapper[4646]: I1203 11:09:46.695956 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b414a389-9e48-412c-b4e0-26a9b4cb51eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b414a389-9e48-412c-b4e0-26a9b4cb51eb" (UID: "b414a389-9e48-412c-b4e0-26a9b4cb51eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:09:46 crc kubenswrapper[4646]: I1203 11:09:46.728407 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b414a389-9e48-412c-b4e0-26a9b4cb51eb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 11:09:46 crc kubenswrapper[4646]: I1203 11:09:46.728447 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b414a389-9e48-412c-b4e0-26a9b4cb51eb-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 11:09:46 crc kubenswrapper[4646]: I1203 11:09:46.728459 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjz45\" (UniqueName: \"kubernetes.io/projected/b414a389-9e48-412c-b4e0-26a9b4cb51eb-kube-api-access-hjz45\") on node \"crc\" DevicePath \"\"" Dec 03 11:09:47 crc kubenswrapper[4646]: I1203 11:09:47.356903 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" event={"ID":"e776d6da-584d-498f-b611-0eb67be6d6d0","Type":"ContainerStarted","Data":"5459ec8dadc0db93a299da8ebbac31cfd0c9ce33f28bbd48d0c165063e6ec014"} Dec 03 11:09:47 crc kubenswrapper[4646]: I1203 11:09:47.360133 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8pb7" event={"ID":"b414a389-9e48-412c-b4e0-26a9b4cb51eb","Type":"ContainerDied","Data":"5552dde4963815fc95260e1b3fa27fa5397491c451448fc32268c147176942aa"} Dec 03 11:09:47 crc kubenswrapper[4646]: I1203 11:09:47.360209 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8pb7" Dec 03 11:09:47 crc kubenswrapper[4646]: I1203 11:09:47.395490 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n8pb7"] Dec 03 11:09:47 crc kubenswrapper[4646]: I1203 11:09:47.401192 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n8pb7"] Dec 03 11:09:47 crc kubenswrapper[4646]: E1203 11:09:47.542023 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 03 11:09:47 crc kubenswrapper[4646]: E1203 11:09:47.542207 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j4bk6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-g8pg8_openstack-operators(9a1416d1-243f-4ab9-9a9c-4b71e0f23876): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:09:47 crc kubenswrapper[4646]: I1203 11:09:47.788914 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp"] Dec 03 11:09:47 crc kubenswrapper[4646]: I1203 11:09:47.872172 4646 scope.go:117] "RemoveContainer" containerID="c559b9c19ff8f4aac03d7c72ac0736c992d01dcf395d05fc3d3fd9d978f225f4" Dec 03 11:09:47 crc kubenswrapper[4646]: W1203 11:09:47.903286 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5dfec9c_60de_45c2_9e93_d3f9ef6fabb0.slice/crio-4f50a18feccfe51dbc50d6940189a795b4e105de0e16c70c797f33c63e3a80cf WatchSource:0}: Error finding container 4f50a18feccfe51dbc50d6940189a795b4e105de0e16c70c797f33c63e3a80cf: Status 404 returned error can't find the container with id 4f50a18feccfe51dbc50d6940189a795b4e105de0e16c70c797f33c63e3a80cf Dec 03 11:09:47 crc kubenswrapper[4646]: I1203 11:09:47.907504 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75816a6e-5fea-4212-964b-aafb2dce1385" path="/var/lib/kubelet/pods/75816a6e-5fea-4212-964b-aafb2dce1385/volumes" Dec 03 11:09:47 crc kubenswrapper[4646]: I1203 11:09:47.908202 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b414a389-9e48-412c-b4e0-26a9b4cb51eb" path="/var/lib/kubelet/pods/b414a389-9e48-412c-b4e0-26a9b4cb51eb/volumes" Dec 03 11:09:47 crc kubenswrapper[4646]: E1203 11:09:47.912249 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 11:09:47 crc kubenswrapper[4646]: E1203 11:09:47.912499 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-npqn5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-schbl_openstack-operators(31371552-226a-436f-a595-52bc387ed5cf): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 11:09:47 crc kubenswrapper[4646]: E1203 11:09:47.913708 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-schbl" podUID="31371552-226a-436f-a595-52bc387ed5cf" Dec 03 11:09:48 crc kubenswrapper[4646]: E1203 11:09:48.049565 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 11:09:48 crc kubenswrapper[4646]: E1203 11:09:48.049725 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wmnwd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-rcmlq_openstack-operators(f3e07f77-c87f-4c6b-be4e-938d71354f8b): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 11:09:48 crc kubenswrapper[4646]: E1203 11:09:48.050847 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rcmlq" podUID="f3e07f77-c87f-4c6b-be4e-938d71354f8b" Dec 03 11:09:48 crc kubenswrapper[4646]: E1203 11:09:48.166088 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 11:09:48 crc kubenswrapper[4646]: E1203 11:09:48.166501 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-st64w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-mzlrq_openstack-operators(396bd18c-37f7-471f-a0f9-0b9cb5a447e3): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 11:09:48 crc kubenswrapper[4646]: E1203 11:09:48.167825 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mzlrq" podUID="396bd18c-37f7-471f-a0f9-0b9cb5a447e3" Dec 03 11:09:48 crc kubenswrapper[4646]: I1203 11:09:48.207240 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7"] Dec 03 11:09:48 crc kubenswrapper[4646]: I1203 11:09:48.369551 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" event={"ID":"d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0","Type":"ContainerStarted","Data":"4f50a18feccfe51dbc50d6940189a795b4e105de0e16c70c797f33c63e3a80cf"} Dec 03 11:09:48 crc kubenswrapper[4646]: I1203 11:09:48.386579 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" event={"ID":"83f119ed-6576-4e47-82b1-a06189cff0a7","Type":"ContainerStarted","Data":"b007c4ab40234b0808cf970a070557254e3989b9c8740959ab018a6431067b4f"} Dec 03 11:09:48 crc kubenswrapper[4646]: I1203 11:09:48.539920 4646 scope.go:117] "RemoveContainer" containerID="9bef189530ac170798df1035eadceeb84e109e1aaf986c471eadf95f2ec2930b" Dec 03 11:09:48 crc kubenswrapper[4646]: I1203 11:09:48.820107 4646 scope.go:117] "RemoveContainer" containerID="489acfe2db031321c437c8680630e180c844e3616468958da33b16fa88ae8053" Dec 03 11:09:49 crc kubenswrapper[4646]: I1203 11:09:49.406313 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hlwj7" event={"ID":"e2fc2280-6814-4ba0-bd48-2272d37f2b41","Type":"ContainerStarted","Data":"9a175ca7312d125042caf9770939efb6509de10b51000910de3f1140decb6ab8"} Dec 03 11:09:52 crc kubenswrapper[4646]: I1203 11:09:52.443887 4646 scope.go:117] "RemoveContainer" containerID="b0c2880f23237eb03391a50ab85c8c308f9d883e054bfe555390c47e792535fe" Dec 03 11:09:52 crc kubenswrapper[4646]: E1203 11:09:52.877211 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 11:09:52 crc kubenswrapper[4646]: E1203 11:09:52.877387 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-h6g6n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-w5c45_openstack-operators(55e5ed61-839c-441b-acf2-cd43cc0c5c90): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 11:09:52 crc kubenswrapper[4646]: E1203 11:09:52.878725 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-w5c45" podUID="55e5ed61-839c-441b-acf2-cd43cc0c5c90" Dec 03 11:09:53 crc kubenswrapper[4646]: I1203 11:09:53.443695 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" event={"ID":"83f119ed-6576-4e47-82b1-a06189cff0a7","Type":"ContainerStarted","Data":"2a58c8c7dbb1bcdec0a88fda59dd27a807a80e2da3f386640c2009ee7b7f2afc"} Dec 03 11:09:53 crc kubenswrapper[4646]: I1203 11:09:53.443742 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:09:53 crc kubenswrapper[4646]: I1203 11:09:53.496677 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" podStartSLOduration=55.496657485 podStartE2EDuration="55.496657485s" podCreationTimestamp="2025-12-03 11:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:09:53.493836943 +0000 UTC m=+969.956893078" watchObservedRunningTime="2025-12-03 11:09:53.496657485 +0000 UTC m=+969.959713620" Dec 03 11:09:54 crc kubenswrapper[4646]: I1203 11:09:54.175377 4646 scope.go:117] "RemoveContainer" containerID="4ba9e5f273af01fd080289ee8fc517db397dd7764f78d9c66529be681a072676" Dec 03 11:09:54 crc kubenswrapper[4646]: E1203 11:09:54.302832 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 11:09:54 crc kubenswrapper[4646]: E1203 11:09:54.302993 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9qjsn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-sc2vn_openstack-operators(eaefadf5-75cd-4ff2-811c-e25bb2850173): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" logger="UnhandledError" Dec 03 11:09:54 crc kubenswrapper[4646]: E1203 11:09:54.304209 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \\\"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\\\": context canceled\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sc2vn" podUID="eaefadf5-75cd-4ff2-811c-e25bb2850173" Dec 03 11:09:56 crc kubenswrapper[4646]: I1203 11:09:56.164191 4646 scope.go:117] "RemoveContainer" containerID="430d69cae0f9e978077e5affa01ee4bc0866f1712455d182c63f9b00c123e2c3" Dec 03 11:09:56 crc kubenswrapper[4646]: I1203 11:09:56.356962 4646 scope.go:117] "RemoveContainer" containerID="e334477b13bc7ef7e1b3512a4a9b40bc77322c4da50bb327cf3228fcfc46cac4" Dec 03 11:09:57 crc kubenswrapper[4646]: E1203 11:09:57.498191 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" podUID="4cb5b92a-540a-41d9-a662-b1f3faf07829" Dec 03 11:09:57 crc kubenswrapper[4646]: E1203 11:09:57.498593 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" podUID="1fed03a1-2388-470c-843e-4de04fd6d9bc" Dec 03 11:09:57 crc kubenswrapper[4646]: E1203 11:09:57.499811 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" podUID="d7ed2042-fe69-40f6-8440-49d63e851ffc" Dec 03 11:09:57 crc kubenswrapper[4646]: E1203 11:09:57.499920 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-l7v6k" podUID="68e6047b-627f-4c45-aa55-4aeab90e17b5" Dec 03 11:09:57 crc kubenswrapper[4646]: E1203 11:09:57.503697 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-t7fj8" podUID="153f2bfa-6ca2-4a9f-900e-5321c5791639" Dec 03 11:09:57 crc kubenswrapper[4646]: E1203 11:09:57.503718 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-f7pcq" podUID="011cd55c-5e76-4dd3-905b-62c587119096" Dec 03 11:09:57 crc kubenswrapper[4646]: E1203 11:09:57.503801 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" podUID="e400aa20-0649-47de-8f10-da6090800eb3" Dec 03 11:09:57 crc kubenswrapper[4646]: E1203 11:09:57.503843 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-92vgw" podUID="a0fa06f6-534e-4ad4-806d-c13cad106809" Dec 03 11:09:57 crc kubenswrapper[4646]: E1203 11:09:57.503895 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" podUID="ed761af7-289e-4f6c-a7b9-710886250a3a" Dec 03 11:09:57 crc kubenswrapper[4646]: E1203 11:09:57.503943 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6hl22" podUID="089f6f35-ab06-4b31-a331-16411a9783e7" Dec 03 11:09:57 crc kubenswrapper[4646]: E1203 11:09:57.504079 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" podUID="9a1416d1-243f-4ab9-9a9c-4b71e0f23876" Dec 03 11:09:57 crc kubenswrapper[4646]: I1203 11:09:57.517316 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sc2vn" event={"ID":"eaefadf5-75cd-4ff2-811c-e25bb2850173","Type":"ContainerStarted","Data":"adb2d13977eceb1156f079ce62f605d0cf72ba688c3c792efeef3600d0ee187b"} Dec 03 11:09:57 crc kubenswrapper[4646]: E1203 11:09:57.531781 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-j2bwb" podUID="f274e7e0-a406-4617-95bc-fbca7ef39526" Dec 03 11:09:57 crc kubenswrapper[4646]: I1203 11:09:57.532056 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-schbl" event={"ID":"31371552-226a-436f-a595-52bc387ed5cf","Type":"ContainerStarted","Data":"4331b8c08dec38f066b90959aa8ca08dd02d9f662c182b7fd55c9e0ae0f2643b"} Dec 03 11:09:57 crc kubenswrapper[4646]: E1203 11:09:57.533558 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6m86" podUID="0be433f8-aae1-48a4-8853-de55ae48f836" Dec 03 11:09:57 crc kubenswrapper[4646]: I1203 11:09:57.542134 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6hl22" event={"ID":"089f6f35-ab06-4b31-a331-16411a9783e7","Type":"ContainerStarted","Data":"2a8bf199013b66e601150e6fb58eb4e6a437fc2f5d660b1a6967eafdbf60ac1a"} Dec 03 11:09:57 crc kubenswrapper[4646]: I1203 11:09:57.674622 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hlwj7" event={"ID":"e2fc2280-6814-4ba0-bd48-2272d37f2b41","Type":"ContainerStarted","Data":"97c6c493824079c5cc80155464a1db9ab26098f7dd32434618e112cba0446e6e"} Dec 03 11:09:57 crc kubenswrapper[4646]: I1203 11:09:57.676044 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hlwj7" Dec 03 11:09:57 crc kubenswrapper[4646]: I1203 11:09:57.681291 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" event={"ID":"d7ed2042-fe69-40f6-8440-49d63e851ffc","Type":"ContainerStarted","Data":"2650169dca4b432c691175c80cccd39952f9ccc401e2d1a515a749666a1b3783"} Dec 03 11:09:57 crc kubenswrapper[4646]: E1203 11:09:57.701105 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" podUID="d7ed2042-fe69-40f6-8440-49d63e851ffc" Dec 03 11:09:57 crc kubenswrapper[4646]: I1203 11:09:57.701437 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hlwj7" Dec 03 11:09:57 crc kubenswrapper[4646]: I1203 11:09:57.721925 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-l7v6k" event={"ID":"68e6047b-627f-4c45-aa55-4aeab90e17b5","Type":"ContainerStarted","Data":"9520f584e21253e57345462636e245eb819dcaac75f632483b90b91ff8617fd8"} Dec 03 11:09:57 crc kubenswrapper[4646]: I1203 11:09:57.728964 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hlwj7" podStartSLOduration=7.16198427 podStartE2EDuration="59.72893958s" podCreationTimestamp="2025-12-03 11:08:58 +0000 UTC" firstStartedPulling="2025-12-03 11:09:03.919714683 +0000 UTC m=+920.382770818" lastFinishedPulling="2025-12-03 11:09:56.486669993 +0000 UTC m=+972.949726128" observedRunningTime="2025-12-03 11:09:57.717815208 +0000 UTC m=+974.180871343" watchObservedRunningTime="2025-12-03 11:09:57.72893958 +0000 UTC m=+974.191995735" Dec 03 11:09:57 crc kubenswrapper[4646]: I1203 11:09:57.756882 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-w5c45" event={"ID":"55e5ed61-839c-441b-acf2-cd43cc0c5c90","Type":"ContainerStarted","Data":"e7c445a0a70e86564663772ecb95697fd65b5269cdbf82c03df211b2581bc37e"} Dec 03 11:09:57 crc kubenswrapper[4646]: I1203 11:09:57.840617 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-f7pcq" event={"ID":"011cd55c-5e76-4dd3-905b-62c587119096","Type":"ContainerStarted","Data":"6e06c67b6d351f364925810cc9bdab1d81faba80751bc6912cd0bc0e4306671b"} Dec 03 11:09:57 crc kubenswrapper[4646]: I1203 11:09:57.894637 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rcmlq" event={"ID":"f3e07f77-c87f-4c6b-be4e-938d71354f8b","Type":"ContainerStarted","Data":"f8b45b757e38f116b3ffb347bd487b99cf7a2decec7e5986a0b3759131f7bf12"} Dec 03 11:09:57 crc kubenswrapper[4646]: I1203 11:09:57.989661 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5vdh" event={"ID":"afd2b4e9-59a4-4e52-bd72-706b8edbd719","Type":"ContainerStarted","Data":"8047b22440752cbb73ffa0fa0b34577f55f2c5a65339946e8452c694eea14dc3"} Dec 03 11:09:58 crc kubenswrapper[4646]: I1203 11:09:58.015643 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" event={"ID":"e776d6da-584d-498f-b611-0eb67be6d6d0","Type":"ContainerStarted","Data":"ef6a4b5ac986fd1a6dae9e60acae8fe621cab1465822fe88e1207cea395dc51a"} Dec 03 11:09:58 crc kubenswrapper[4646]: I1203 11:09:58.051070 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-b5vdh" podStartSLOduration=15.270100575 podStartE2EDuration="1m0.051052713s" podCreationTimestamp="2025-12-03 11:08:58 +0000 UTC" firstStartedPulling="2025-12-03 11:09:04.370769864 +0000 UTC m=+920.833825989" lastFinishedPulling="2025-12-03 11:09:49.151721992 +0000 UTC m=+965.614778127" observedRunningTime="2025-12-03 11:09:58.050732324 +0000 UTC m=+974.513788459" watchObservedRunningTime="2025-12-03 11:09:58.051052713 +0000 UTC m=+974.514108848" Dec 03 11:09:58 crc kubenswrapper[4646]: I1203 11:09:58.063622 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-t7fj8" event={"ID":"153f2bfa-6ca2-4a9f-900e-5321c5791639","Type":"ContainerStarted","Data":"05a58f8207cb34c128b8b40fb23d4d64d16936be62e298e681ce2fc69961ee5f"} Dec 03 11:09:58 crc kubenswrapper[4646]: I1203 11:09:58.088586 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" event={"ID":"1fed03a1-2388-470c-843e-4de04fd6d9bc","Type":"ContainerStarted","Data":"77235e4e091188533c08023545eeed36549702b868f9c1954217f57f76e6cd3b"} Dec 03 11:09:58 crc kubenswrapper[4646]: E1203 11:09:58.097631 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809\\\"\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" podUID="1fed03a1-2388-470c-843e-4de04fd6d9bc" Dec 03 11:09:58 crc kubenswrapper[4646]: I1203 11:09:58.110239 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" event={"ID":"4cb5b92a-540a-41d9-a662-b1f3faf07829","Type":"ContainerStarted","Data":"b94e1d8318f5f98eb5349eb869b42067992eacf81a115634739736bd13318264"} Dec 03 11:09:58 crc kubenswrapper[4646]: E1203 11:09:58.118655 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85\\\"\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" podUID="4cb5b92a-540a-41d9-a662-b1f3faf07829" Dec 03 11:09:58 crc kubenswrapper[4646]: I1203 11:09:58.135960 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" event={"ID":"e400aa20-0649-47de-8f10-da6090800eb3","Type":"ContainerStarted","Data":"451cc7ccf57559257c83a7bd31afc702644510d1b136c5d341511d6c8aafcaf1"} Dec 03 11:09:58 crc kubenswrapper[4646]: E1203 11:09:58.139626 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" podUID="e400aa20-0649-47de-8f10-da6090800eb3" Dec 03 11:09:58 crc kubenswrapper[4646]: I1203 11:09:58.158934 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" event={"ID":"9a1416d1-243f-4ab9-9a9c-4b71e0f23876","Type":"ContainerStarted","Data":"07478284a50cd1f0150b5ac08f16679cb6ee1c7a96c8bd7edaa84805732b5b2d"} Dec 03 11:09:58 crc kubenswrapper[4646]: E1203 11:09:58.170505 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" podUID="9a1416d1-243f-4ab9-9a9c-4b71e0f23876" Dec 03 11:09:58 crc kubenswrapper[4646]: I1203 11:09:58.182852 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" event={"ID":"ed761af7-289e-4f6c-a7b9-710886250a3a","Type":"ContainerStarted","Data":"3c21d2c13a5ff705070a9a264372c2fda0e8b51fd266e726b27eef20613f8308"} Dec 03 11:09:58 crc kubenswrapper[4646]: E1203 11:09:58.184092 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" podUID="ed761af7-289e-4f6c-a7b9-710886250a3a" Dec 03 11:09:58 crc kubenswrapper[4646]: I1203 11:09:58.193198 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mzlrq" event={"ID":"396bd18c-37f7-471f-a0f9-0b9cb5a447e3","Type":"ContainerStarted","Data":"7e3612d17e255c1b90df43b11075d7976165564228f4ca4300bdf312edbd2fa7"} Dec 03 11:09:58 crc kubenswrapper[4646]: I1203 11:09:58.193985 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mzlrq" Dec 03 11:09:58 crc kubenswrapper[4646]: I1203 11:09:58.219058 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" event={"ID":"d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0","Type":"ContainerStarted","Data":"0ce8a0e961bf563e3f5189e2ec753cfa98e13df9a81b39db590177328530e532"} Dec 03 11:09:58 crc kubenswrapper[4646]: I1203 11:09:58.219711 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" Dec 03 11:09:58 crc kubenswrapper[4646]: I1203 11:09:58.258026 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-92vgw" event={"ID":"a0fa06f6-534e-4ad4-806d-c13cad106809","Type":"ContainerStarted","Data":"71a45ca604a362f372f8e5a9c99c64dbd261fd18c9e900335cdfb68084baf02d"} Dec 03 11:09:58 crc kubenswrapper[4646]: I1203 11:09:58.400506 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mzlrq" podStartSLOduration=9.321879495 podStartE2EDuration="1m1.400486396s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:04.049517786 +0000 UTC m=+920.512573921" lastFinishedPulling="2025-12-03 11:09:56.128124687 +0000 UTC m=+972.591180822" observedRunningTime="2025-12-03 11:09:58.367833522 +0000 UTC m=+974.830889657" watchObservedRunningTime="2025-12-03 11:09:58.400486396 +0000 UTC m=+974.863542531" Dec 03 11:09:58 crc kubenswrapper[4646]: I1203 11:09:58.451441 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" podStartSLOduration=53.112213991 podStartE2EDuration="1m1.451401988s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:47.909875657 +0000 UTC m=+964.372931792" lastFinishedPulling="2025-12-03 11:09:56.249063664 +0000 UTC m=+972.712119789" observedRunningTime="2025-12-03 11:09:58.449499733 +0000 UTC m=+974.912555868" watchObservedRunningTime="2025-12-03 11:09:58.451401988 +0000 UTC m=+974.914458123" Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.265090 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-j2bwb" event={"ID":"f274e7e0-a406-4617-95bc-fbca7ef39526","Type":"ContainerStarted","Data":"7754f700ef9a581da944a62a304552ae20d0ea6fd15dbbc9eda107139ab5a670"} Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.266606 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sc2vn" event={"ID":"eaefadf5-75cd-4ff2-811c-e25bb2850173","Type":"ContainerStarted","Data":"795f416069f73538b46cf7f015e56718bfed99121f8ac46a8bcb48a600197bd8"} Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.267430 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sc2vn" Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.271106 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-schbl" event={"ID":"31371552-226a-436f-a595-52bc387ed5cf","Type":"ContainerStarted","Data":"1d44e71c980515c33bc70c94046f6d592c300e4363e97850809f58eb5bd8c244"} Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.271514 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-schbl" Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.272931 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-w5c45" event={"ID":"55e5ed61-839c-441b-acf2-cd43cc0c5c90","Type":"ContainerStarted","Data":"04e7c6f2544d95dc78872882faefe02853361755c473cdf95ae5d3502674a6e8"} Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.272986 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-w5c45" Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.274198 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" event={"ID":"d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0","Type":"ContainerStarted","Data":"64dc522d7b86c31de5e9aa55fc2b7f647d18741a55051917ad2908f66831489b"} Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.275889 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mzlrq" event={"ID":"396bd18c-37f7-471f-a0f9-0b9cb5a447e3","Type":"ContainerStarted","Data":"80b1cb7a52d776b6966287b91f6e0435afcdba682b85b1bf63da7b941c782f45"} Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.277650 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rcmlq" event={"ID":"f3e07f77-c87f-4c6b-be4e-938d71354f8b","Type":"ContainerStarted","Data":"07e4d0b5145402c85f7555c3ba79a15facaa8e94e4b52a010d3ae8b0c5316cdd"} Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.277984 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rcmlq" Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.279105 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" event={"ID":"e776d6da-584d-498f-b611-0eb67be6d6d0","Type":"ContainerStarted","Data":"69622f993c0416905df83834b192358322184096034209142b748f099276a853"} Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.279478 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.285193 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6m86" event={"ID":"0be433f8-aae1-48a4-8853-de55ae48f836","Type":"ContainerStarted","Data":"c3f67801124b82a9b90e84c799ad5dcb797aac9e36e06d1129631700716c3d11"} Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.285223 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6m86" event={"ID":"0be433f8-aae1-48a4-8853-de55ae48f836","Type":"ContainerStarted","Data":"f81525991b2565c5686b70e7969a213a961f89299e3da2eab22ae0ef772fe4d1"} Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.322972 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sc2vn" podStartSLOduration=9.731771326 podStartE2EDuration="1m2.322958397s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:03.758944795 +0000 UTC m=+920.222000920" lastFinishedPulling="2025-12-03 11:09:56.350131866 +0000 UTC m=+972.813187991" observedRunningTime="2025-12-03 11:09:59.315754349 +0000 UTC m=+975.778810484" watchObservedRunningTime="2025-12-03 11:09:59.322958397 +0000 UTC m=+975.786014532" Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.389525 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-schbl" podStartSLOduration=9.974547437 podStartE2EDuration="1m2.389510152s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:03.747580067 +0000 UTC m=+920.210636202" lastFinishedPulling="2025-12-03 11:09:56.162542782 +0000 UTC m=+972.625598917" observedRunningTime="2025-12-03 11:09:59.347147757 +0000 UTC m=+975.810203892" watchObservedRunningTime="2025-12-03 11:09:59.389510152 +0000 UTC m=+975.852566287" Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.389655 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" podStartSLOduration=52.542296155 podStartE2EDuration="1m2.389651146s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:46.385220046 +0000 UTC m=+962.848276181" lastFinishedPulling="2025-12-03 11:09:56.232575037 +0000 UTC m=+972.695631172" observedRunningTime="2025-12-03 11:09:59.387671068 +0000 UTC m=+975.850727203" watchObservedRunningTime="2025-12-03 11:09:59.389651146 +0000 UTC m=+975.852707281" Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.450599 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-w5c45" podStartSLOduration=8.197423797 podStartE2EDuration="1m1.450549526s" podCreationTimestamp="2025-12-03 11:08:58 +0000 UTC" firstStartedPulling="2025-12-03 11:09:03.097045328 +0000 UTC m=+919.560101463" lastFinishedPulling="2025-12-03 11:09:56.350171057 +0000 UTC m=+972.813227192" observedRunningTime="2025-12-03 11:09:59.444922604 +0000 UTC m=+975.907978739" watchObservedRunningTime="2025-12-03 11:09:59.450549526 +0000 UTC m=+975.913605661" Dec 03 11:09:59 crc kubenswrapper[4646]: I1203 11:09:59.491310 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rcmlq" podStartSLOduration=10.018190577 podStartE2EDuration="1m2.491290184s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:03.759262904 +0000 UTC m=+920.222319039" lastFinishedPulling="2025-12-03 11:09:56.232362511 +0000 UTC m=+972.695418646" observedRunningTime="2025-12-03 11:09:59.489858923 +0000 UTC m=+975.952915068" watchObservedRunningTime="2025-12-03 11:09:59.491290184 +0000 UTC m=+975.954346309" Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.311144 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-l7v6k" event={"ID":"68e6047b-627f-4c45-aa55-4aeab90e17b5","Type":"ContainerStarted","Data":"8a8750ca8b901a01d0b3c5aee4225328786bab000d0683f567ac629fe1158d4a"} Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.311519 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-l7v6k" Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.313552 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6hl22" event={"ID":"089f6f35-ab06-4b31-a331-16411a9783e7","Type":"ContainerStarted","Data":"cc70b8080002148afe78ff126a5eacf0136a4dcadfda5165257bd3051b716252"} Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.314314 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6hl22" Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.316774 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-t7fj8" event={"ID":"153f2bfa-6ca2-4a9f-900e-5321c5791639","Type":"ContainerStarted","Data":"5163ec4fdabae4810127f501aef0739352627a7ef86a9b036a4152dd61215a5a"} Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.316865 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-t7fj8" Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.320547 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-92vgw" event={"ID":"a0fa06f6-534e-4ad4-806d-c13cad106809","Type":"ContainerStarted","Data":"39e438ba3cd823d7b0b925a5320d4b5e6b42cbce5b86b4cd7bfeb8d19ab786c4"} Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.320674 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-92vgw" Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.322245 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-j2bwb" event={"ID":"f274e7e0-a406-4617-95bc-fbca7ef39526","Type":"ContainerStarted","Data":"a10b9d0ff2bf07cb9dce3df63faea0290cf6ea4666b4be39675be9a57fd15464"} Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.322315 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-j2bwb" Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.324180 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-f7pcq" event={"ID":"011cd55c-5e76-4dd3-905b-62c587119096","Type":"ContainerStarted","Data":"9d1cadf613bc85c38c92a617d86371ecac07eaa205f08233a25e904f8329758b"} Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.339117 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-l7v6k" podStartSLOduration=7.652079958 podStartE2EDuration="1m3.339099967s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:02.944801876 +0000 UTC m=+919.407858011" lastFinishedPulling="2025-12-03 11:09:58.631821885 +0000 UTC m=+975.094878020" observedRunningTime="2025-12-03 11:10:00.338081847 +0000 UTC m=+976.801137982" watchObservedRunningTime="2025-12-03 11:10:00.339099967 +0000 UTC m=+976.802156102" Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.397619 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-f7pcq" podStartSLOduration=7.565287949 podStartE2EDuration="1m3.397601528s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:02.941881072 +0000 UTC m=+919.404937207" lastFinishedPulling="2025-12-03 11:09:58.774194651 +0000 UTC m=+975.237250786" observedRunningTime="2025-12-03 11:10:00.369841716 +0000 UTC m=+976.832897851" watchObservedRunningTime="2025-12-03 11:10:00.397601528 +0000 UTC m=+976.860657663" Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.441489 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-t7fj8" podStartSLOduration=8.670223285 podStartE2EDuration="1m3.441468177s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:03.999314374 +0000 UTC m=+920.462370509" lastFinishedPulling="2025-12-03 11:09:58.770559266 +0000 UTC m=+975.233615401" observedRunningTime="2025-12-03 11:10:00.405222239 +0000 UTC m=+976.868278374" watchObservedRunningTime="2025-12-03 11:10:00.441468177 +0000 UTC m=+976.904524312" Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.443704 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-j2bwb" podStartSLOduration=8.980252699 podStartE2EDuration="1m3.443696971s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:04.02440844 +0000 UTC m=+920.487464575" lastFinishedPulling="2025-12-03 11:09:58.487852712 +0000 UTC m=+974.950908847" observedRunningTime="2025-12-03 11:10:00.438018577 +0000 UTC m=+976.901074712" watchObservedRunningTime="2025-12-03 11:10:00.443696971 +0000 UTC m=+976.906753106" Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.480455 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6m86" podStartSLOduration=9.094996356 podStartE2EDuration="1m3.480438123s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:04.101519389 +0000 UTC m=+920.564575534" lastFinishedPulling="2025-12-03 11:09:58.486961176 +0000 UTC m=+974.950017301" observedRunningTime="2025-12-03 11:10:00.472205055 +0000 UTC m=+976.935261190" watchObservedRunningTime="2025-12-03 11:10:00.480438123 +0000 UTC m=+976.943494248" Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.503476 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6c5c989645-kkkb7" Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.542503 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6hl22" podStartSLOduration=8.95471422 podStartE2EDuration="1m3.542479867s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:03.89334748 +0000 UTC m=+920.356403615" lastFinishedPulling="2025-12-03 11:09:58.481113127 +0000 UTC m=+974.944169262" observedRunningTime="2025-12-03 11:10:00.537671998 +0000 UTC m=+977.000728133" watchObservedRunningTime="2025-12-03 11:10:00.542479867 +0000 UTC m=+977.005535992" Dec 03 11:10:00 crc kubenswrapper[4646]: I1203 11:10:00.568536 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-92vgw" podStartSLOduration=7.746426045 podStartE2EDuration="1m3.5685183s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:04.157745635 +0000 UTC m=+920.620801790" lastFinishedPulling="2025-12-03 11:09:59.97983791 +0000 UTC m=+976.442894045" observedRunningTime="2025-12-03 11:10:00.565630036 +0000 UTC m=+977.028686171" watchObservedRunningTime="2025-12-03 11:10:00.5685183 +0000 UTC m=+977.031574435" Dec 03 11:10:01 crc kubenswrapper[4646]: I1203 11:10:01.332119 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-f7pcq" Dec 03 11:10:01 crc kubenswrapper[4646]: I1203 11:10:01.333796 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-schbl" Dec 03 11:10:03 crc kubenswrapper[4646]: I1203 11:10:03.565744 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6jbp" Dec 03 11:10:03 crc kubenswrapper[4646]: I1203 11:10:03.914439 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9" Dec 03 11:10:07 crc kubenswrapper[4646]: I1203 11:10:07.859654 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mzlrq" Dec 03 11:10:07 crc kubenswrapper[4646]: I1203 11:10:07.914739 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6hl22" Dec 03 11:10:08 crc kubenswrapper[4646]: I1203 11:10:08.136677 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-l7v6k" Dec 03 11:10:08 crc kubenswrapper[4646]: I1203 11:10:08.146199 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rcmlq" Dec 03 11:10:08 crc kubenswrapper[4646]: I1203 11:10:08.165260 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-92vgw" Dec 03 11:10:08 crc kubenswrapper[4646]: I1203 11:10:08.254696 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sc2vn" Dec 03 11:10:08 crc kubenswrapper[4646]: I1203 11:10:08.322681 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-t7fj8" Dec 03 11:10:08 crc kubenswrapper[4646]: I1203 11:10:08.439548 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-f7pcq" Dec 03 11:10:08 crc kubenswrapper[4646]: I1203 11:10:08.443437 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-j2bwb" Dec 03 11:10:08 crc kubenswrapper[4646]: I1203 11:10:08.521186 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-w5c45" Dec 03 11:10:08 crc kubenswrapper[4646]: I1203 11:10:08.546444 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6m86" Dec 03 11:10:08 crc kubenswrapper[4646]: I1203 11:10:08.548420 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z6m86" Dec 03 11:10:10 crc kubenswrapper[4646]: I1203 11:10:10.397419 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" event={"ID":"4cb5b92a-540a-41d9-a662-b1f3faf07829","Type":"ContainerStarted","Data":"0849e629a6e89eb2d8156b2ec0d76ebe766c933921c8f2d3f47d0f5625dc201a"} Dec 03 11:10:10 crc kubenswrapper[4646]: I1203 11:10:10.398100 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" Dec 03 11:10:10 crc kubenswrapper[4646]: I1203 11:10:10.420879 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" podStartSLOduration=8.322398388 podStartE2EDuration="1m13.420859085s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:04.272137982 +0000 UTC m=+920.735194117" lastFinishedPulling="2025-12-03 11:10:09.370598679 +0000 UTC m=+985.833654814" observedRunningTime="2025-12-03 11:10:10.416287733 +0000 UTC m=+986.879343888" watchObservedRunningTime="2025-12-03 11:10:10.420859085 +0000 UTC m=+986.883915220" Dec 03 11:10:11 crc kubenswrapper[4646]: I1203 11:10:11.406611 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" event={"ID":"d7ed2042-fe69-40f6-8440-49d63e851ffc","Type":"ContainerStarted","Data":"a6dd8c3b3124c1034517e4c6ff894c919245c09bc2e6444d0eb33e6332c935b9"} Dec 03 11:10:11 crc kubenswrapper[4646]: I1203 11:10:11.406874 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" Dec 03 11:10:11 crc kubenswrapper[4646]: I1203 11:10:11.408401 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" event={"ID":"ed761af7-289e-4f6c-a7b9-710886250a3a","Type":"ContainerStarted","Data":"efb92fc4dcfae41f4e1da2df01751897472cd3fa15b32098ab84ea1d4cb98126"} Dec 03 11:10:11 crc kubenswrapper[4646]: I1203 11:10:11.408718 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" Dec 03 11:10:11 crc kubenswrapper[4646]: I1203 11:10:11.429354 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" podStartSLOduration=8.116161455 podStartE2EDuration="1m14.429323672s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:04.258541329 +0000 UTC m=+920.721597454" lastFinishedPulling="2025-12-03 11:10:10.571703526 +0000 UTC m=+987.034759671" observedRunningTime="2025-12-03 11:10:11.425297216 +0000 UTC m=+987.888353351" watchObservedRunningTime="2025-12-03 11:10:11.429323672 +0000 UTC m=+987.892379807" Dec 03 11:10:11 crc kubenswrapper[4646]: I1203 11:10:11.457086 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" podStartSLOduration=8.047568481 podStartE2EDuration="1m14.457065354s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:04.375110019 +0000 UTC m=+920.838166154" lastFinishedPulling="2025-12-03 11:10:10.784606892 +0000 UTC m=+987.247663027" observedRunningTime="2025-12-03 11:10:11.447409035 +0000 UTC m=+987.910465190" watchObservedRunningTime="2025-12-03 11:10:11.457065354 +0000 UTC m=+987.920121489" Dec 03 11:10:12 crc kubenswrapper[4646]: I1203 11:10:12.417323 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" event={"ID":"1fed03a1-2388-470c-843e-4de04fd6d9bc","Type":"ContainerStarted","Data":"cbf2429ce79a406edcf0754f0398e9327d0960b0ec8d247f8db813f86935c077"} Dec 03 11:10:12 crc kubenswrapper[4646]: I1203 11:10:12.418518 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" Dec 03 11:10:12 crc kubenswrapper[4646]: I1203 11:10:12.444721 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" podStartSLOduration=8.381523986 podStartE2EDuration="1m15.444705469s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:04.2917877 +0000 UTC m=+920.754843835" lastFinishedPulling="2025-12-03 11:10:11.354969183 +0000 UTC m=+987.818025318" observedRunningTime="2025-12-03 11:10:12.440043574 +0000 UTC m=+988.903099709" watchObservedRunningTime="2025-12-03 11:10:12.444705469 +0000 UTC m=+988.907761604" Dec 03 11:10:14 crc kubenswrapper[4646]: I1203 11:10:14.430516 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" event={"ID":"e400aa20-0649-47de-8f10-da6090800eb3","Type":"ContainerStarted","Data":"a685cf3a9c8297624266daadd42dd679ca6dc43ce2e2de175de43c24364abafc"} Dec 03 11:10:14 crc kubenswrapper[4646]: I1203 11:10:14.430935 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" Dec 03 11:10:14 crc kubenswrapper[4646]: I1203 11:10:14.432208 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" event={"ID":"9a1416d1-243f-4ab9-9a9c-4b71e0f23876","Type":"ContainerStarted","Data":"d4b7264c39cfdafe8d16dba5a0dad861f37cb3ccd6a3e5db7deb497a6c1301fa"} Dec 03 11:10:14 crc kubenswrapper[4646]: I1203 11:10:14.432497 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" Dec 03 11:10:14 crc kubenswrapper[4646]: I1203 11:10:14.451784 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" podStartSLOduration=8.361518139 podStartE2EDuration="1m17.451767339s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:04.265599293 +0000 UTC m=+920.728655428" lastFinishedPulling="2025-12-03 11:10:13.355848493 +0000 UTC m=+989.818904628" observedRunningTime="2025-12-03 11:10:14.445632571 +0000 UTC m=+990.908688706" watchObservedRunningTime="2025-12-03 11:10:14.451767339 +0000 UTC m=+990.914823474" Dec 03 11:10:14 crc kubenswrapper[4646]: I1203 11:10:14.464501 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" podStartSLOduration=8.379995814 podStartE2EDuration="1m17.464476026s" podCreationTimestamp="2025-12-03 11:08:57 +0000 UTC" firstStartedPulling="2025-12-03 11:09:04.27240223 +0000 UTC m=+920.735458365" lastFinishedPulling="2025-12-03 11:10:13.356882442 +0000 UTC m=+989.819938577" observedRunningTime="2025-12-03 11:10:14.460952934 +0000 UTC m=+990.924009089" watchObservedRunningTime="2025-12-03 11:10:14.464476026 +0000 UTC m=+990.927532161" Dec 03 11:10:17 crc kubenswrapper[4646]: I1203 11:10:17.759273 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pnbk8" Dec 03 11:10:17 crc kubenswrapper[4646]: I1203 11:10:17.792415 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ps2zm" Dec 03 11:10:17 crc kubenswrapper[4646]: I1203 11:10:17.816567 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-56r4n" Dec 03 11:10:18 crc kubenswrapper[4646]: I1203 11:10:18.228693 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-g8pg8" Dec 03 11:10:18 crc kubenswrapper[4646]: I1203 11:10:18.279068 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-n9s99" Dec 03 11:10:25 crc kubenswrapper[4646]: I1203 11:10:25.963888 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:10:25 crc kubenswrapper[4646]: I1203 11:10:25.964240 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:10:27 crc kubenswrapper[4646]: I1203 11:10:27.772055 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ss5tr" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.044529 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-v4gdc"] Dec 03 11:10:42 crc kubenswrapper[4646]: E1203 11:10:42.045205 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b414a389-9e48-412c-b4e0-26a9b4cb51eb" containerName="extract-utilities" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.045217 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="b414a389-9e48-412c-b4e0-26a9b4cb51eb" containerName="extract-utilities" Dec 03 11:10:42 crc kubenswrapper[4646]: E1203 11:10:42.045240 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42a46b9b-c088-4e78-848e-c5b605694d3e" containerName="extract-content" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.045246 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="42a46b9b-c088-4e78-848e-c5b605694d3e" containerName="extract-content" Dec 03 11:10:42 crc kubenswrapper[4646]: E1203 11:10:42.045259 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75816a6e-5fea-4212-964b-aafb2dce1385" containerName="extract-content" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.045265 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="75816a6e-5fea-4212-964b-aafb2dce1385" containerName="extract-content" Dec 03 11:10:42 crc kubenswrapper[4646]: E1203 11:10:42.045277 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42a46b9b-c088-4e78-848e-c5b605694d3e" containerName="registry-server" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.045282 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="42a46b9b-c088-4e78-848e-c5b605694d3e" containerName="registry-server" Dec 03 11:10:42 crc kubenswrapper[4646]: E1203 11:10:42.045300 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42a46b9b-c088-4e78-848e-c5b605694d3e" containerName="extract-utilities" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.045306 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="42a46b9b-c088-4e78-848e-c5b605694d3e" containerName="extract-utilities" Dec 03 11:10:42 crc kubenswrapper[4646]: E1203 11:10:42.045321 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b414a389-9e48-412c-b4e0-26a9b4cb51eb" containerName="registry-server" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.045327 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="b414a389-9e48-412c-b4e0-26a9b4cb51eb" containerName="registry-server" Dec 03 11:10:42 crc kubenswrapper[4646]: E1203 11:10:42.045350 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75816a6e-5fea-4212-964b-aafb2dce1385" containerName="registry-server" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.045355 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="75816a6e-5fea-4212-964b-aafb2dce1385" containerName="registry-server" Dec 03 11:10:42 crc kubenswrapper[4646]: E1203 11:10:42.045368 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b414a389-9e48-412c-b4e0-26a9b4cb51eb" containerName="extract-content" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.045374 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="b414a389-9e48-412c-b4e0-26a9b4cb51eb" containerName="extract-content" Dec 03 11:10:42 crc kubenswrapper[4646]: E1203 11:10:42.045387 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75816a6e-5fea-4212-964b-aafb2dce1385" containerName="extract-utilities" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.045393 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="75816a6e-5fea-4212-964b-aafb2dce1385" containerName="extract-utilities" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.045520 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="42a46b9b-c088-4e78-848e-c5b605694d3e" containerName="registry-server" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.045538 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="75816a6e-5fea-4212-964b-aafb2dce1385" containerName="registry-server" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.045550 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="b414a389-9e48-412c-b4e0-26a9b4cb51eb" containerName="registry-server" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.046320 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-v4gdc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.051183 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.053524 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.053722 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.053784 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-4ppn6" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.059097 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-v4gdc"] Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.146848 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zxtkc"] Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.148034 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.163044 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zxtkc"] Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.165721 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.216824 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b7835bf-ecbc-457c-bf60-86652b766d4e-config\") pod \"dnsmasq-dns-78dd6ddcc-zxtkc\" (UID: \"3b7835bf-ecbc-457c-bf60-86652b766d4e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.216869 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b7835bf-ecbc-457c-bf60-86652b766d4e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zxtkc\" (UID: \"3b7835bf-ecbc-457c-bf60-86652b766d4e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.216899 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cc7fc28-9dc3-4563-a188-542b67825477-config\") pod \"dnsmasq-dns-675f4bcbfc-v4gdc\" (UID: \"8cc7fc28-9dc3-4563-a188-542b67825477\") " pod="openstack/dnsmasq-dns-675f4bcbfc-v4gdc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.216928 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t945f\" (UniqueName: \"kubernetes.io/projected/3b7835bf-ecbc-457c-bf60-86652b766d4e-kube-api-access-t945f\") pod \"dnsmasq-dns-78dd6ddcc-zxtkc\" (UID: \"3b7835bf-ecbc-457c-bf60-86652b766d4e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.216960 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9ptf\" (UniqueName: \"kubernetes.io/projected/8cc7fc28-9dc3-4563-a188-542b67825477-kube-api-access-l9ptf\") pod \"dnsmasq-dns-675f4bcbfc-v4gdc\" (UID: \"8cc7fc28-9dc3-4563-a188-542b67825477\") " pod="openstack/dnsmasq-dns-675f4bcbfc-v4gdc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.318128 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9ptf\" (UniqueName: \"kubernetes.io/projected/8cc7fc28-9dc3-4563-a188-542b67825477-kube-api-access-l9ptf\") pod \"dnsmasq-dns-675f4bcbfc-v4gdc\" (UID: \"8cc7fc28-9dc3-4563-a188-542b67825477\") " pod="openstack/dnsmasq-dns-675f4bcbfc-v4gdc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.318265 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b7835bf-ecbc-457c-bf60-86652b766d4e-config\") pod \"dnsmasq-dns-78dd6ddcc-zxtkc\" (UID: \"3b7835bf-ecbc-457c-bf60-86652b766d4e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.318294 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b7835bf-ecbc-457c-bf60-86652b766d4e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zxtkc\" (UID: \"3b7835bf-ecbc-457c-bf60-86652b766d4e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.318323 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cc7fc28-9dc3-4563-a188-542b67825477-config\") pod \"dnsmasq-dns-675f4bcbfc-v4gdc\" (UID: \"8cc7fc28-9dc3-4563-a188-542b67825477\") " pod="openstack/dnsmasq-dns-675f4bcbfc-v4gdc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.318373 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t945f\" (UniqueName: \"kubernetes.io/projected/3b7835bf-ecbc-457c-bf60-86652b766d4e-kube-api-access-t945f\") pod \"dnsmasq-dns-78dd6ddcc-zxtkc\" (UID: \"3b7835bf-ecbc-457c-bf60-86652b766d4e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.320424 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b7835bf-ecbc-457c-bf60-86652b766d4e-config\") pod \"dnsmasq-dns-78dd6ddcc-zxtkc\" (UID: \"3b7835bf-ecbc-457c-bf60-86652b766d4e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.321070 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b7835bf-ecbc-457c-bf60-86652b766d4e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zxtkc\" (UID: \"3b7835bf-ecbc-457c-bf60-86652b766d4e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.321868 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cc7fc28-9dc3-4563-a188-542b67825477-config\") pod \"dnsmasq-dns-675f4bcbfc-v4gdc\" (UID: \"8cc7fc28-9dc3-4563-a188-542b67825477\") " pod="openstack/dnsmasq-dns-675f4bcbfc-v4gdc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.344148 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t945f\" (UniqueName: \"kubernetes.io/projected/3b7835bf-ecbc-457c-bf60-86652b766d4e-kube-api-access-t945f\") pod \"dnsmasq-dns-78dd6ddcc-zxtkc\" (UID: \"3b7835bf-ecbc-457c-bf60-86652b766d4e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.344182 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9ptf\" (UniqueName: \"kubernetes.io/projected/8cc7fc28-9dc3-4563-a188-542b67825477-kube-api-access-l9ptf\") pod \"dnsmasq-dns-675f4bcbfc-v4gdc\" (UID: \"8cc7fc28-9dc3-4563-a188-542b67825477\") " pod="openstack/dnsmasq-dns-675f4bcbfc-v4gdc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.363453 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-v4gdc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.464281 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.838222 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-v4gdc"] Dec 03 11:10:42 crc kubenswrapper[4646]: I1203 11:10:42.992895 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zxtkc"] Dec 03 11:10:42 crc kubenswrapper[4646]: W1203 11:10:42.996497 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b7835bf_ecbc_457c_bf60_86652b766d4e.slice/crio-8608997c794865c6c3ec08e66de9c5bd7450af318f91c1d0fa2938153b0fe421 WatchSource:0}: Error finding container 8608997c794865c6c3ec08e66de9c5bd7450af318f91c1d0fa2938153b0fe421: Status 404 returned error can't find the container with id 8608997c794865c6c3ec08e66de9c5bd7450af318f91c1d0fa2938153b0fe421 Dec 03 11:10:43 crc kubenswrapper[4646]: I1203 11:10:43.663771 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-v4gdc" event={"ID":"8cc7fc28-9dc3-4563-a188-542b67825477","Type":"ContainerStarted","Data":"4dddba3416bad17d1fa48e899d2bec16902768498070123fcb349bdb2b6ea31e"} Dec 03 11:10:43 crc kubenswrapper[4646]: I1203 11:10:43.666499 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" event={"ID":"3b7835bf-ecbc-457c-bf60-86652b766d4e","Type":"ContainerStarted","Data":"8608997c794865c6c3ec08e66de9c5bd7450af318f91c1d0fa2938153b0fe421"} Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.277948 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-v4gdc"] Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.313211 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-44kpk"] Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.315165 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-44kpk" Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.336040 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-44kpk"] Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.522058 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-config\") pod \"dnsmasq-dns-666b6646f7-44kpk\" (UID: \"842907e9-4a8c-494c-a7e1-1d0686eaeb7f\") " pod="openstack/dnsmasq-dns-666b6646f7-44kpk" Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.522138 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-dns-svc\") pod \"dnsmasq-dns-666b6646f7-44kpk\" (UID: \"842907e9-4a8c-494c-a7e1-1d0686eaeb7f\") " pod="openstack/dnsmasq-dns-666b6646f7-44kpk" Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.522174 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlnx4\" (UniqueName: \"kubernetes.io/projected/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-kube-api-access-vlnx4\") pod \"dnsmasq-dns-666b6646f7-44kpk\" (UID: \"842907e9-4a8c-494c-a7e1-1d0686eaeb7f\") " pod="openstack/dnsmasq-dns-666b6646f7-44kpk" Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.623914 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-dns-svc\") pod \"dnsmasq-dns-666b6646f7-44kpk\" (UID: \"842907e9-4a8c-494c-a7e1-1d0686eaeb7f\") " pod="openstack/dnsmasq-dns-666b6646f7-44kpk" Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.624007 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlnx4\" (UniqueName: \"kubernetes.io/projected/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-kube-api-access-vlnx4\") pod \"dnsmasq-dns-666b6646f7-44kpk\" (UID: \"842907e9-4a8c-494c-a7e1-1d0686eaeb7f\") " pod="openstack/dnsmasq-dns-666b6646f7-44kpk" Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.624050 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-config\") pod \"dnsmasq-dns-666b6646f7-44kpk\" (UID: \"842907e9-4a8c-494c-a7e1-1d0686eaeb7f\") " pod="openstack/dnsmasq-dns-666b6646f7-44kpk" Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.625197 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-config\") pod \"dnsmasq-dns-666b6646f7-44kpk\" (UID: \"842907e9-4a8c-494c-a7e1-1d0686eaeb7f\") " pod="openstack/dnsmasq-dns-666b6646f7-44kpk" Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.625928 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-dns-svc\") pod \"dnsmasq-dns-666b6646f7-44kpk\" (UID: \"842907e9-4a8c-494c-a7e1-1d0686eaeb7f\") " pod="openstack/dnsmasq-dns-666b6646f7-44kpk" Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.653201 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlnx4\" (UniqueName: \"kubernetes.io/projected/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-kube-api-access-vlnx4\") pod \"dnsmasq-dns-666b6646f7-44kpk\" (UID: \"842907e9-4a8c-494c-a7e1-1d0686eaeb7f\") " pod="openstack/dnsmasq-dns-666b6646f7-44kpk" Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.712744 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zxtkc"] Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.743311 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2lkc7"] Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.747516 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.783844 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2lkc7"] Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.929839 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-config\") pod \"dnsmasq-dns-57d769cc4f-2lkc7\" (UID: \"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.929991 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq64l\" (UniqueName: \"kubernetes.io/projected/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-kube-api-access-jq64l\") pod \"dnsmasq-dns-57d769cc4f-2lkc7\" (UID: \"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.930042 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-2lkc7\" (UID: \"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" Dec 03 11:10:45 crc kubenswrapper[4646]: I1203 11:10:45.947839 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-44kpk" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.034671 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq64l\" (UniqueName: \"kubernetes.io/projected/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-kube-api-access-jq64l\") pod \"dnsmasq-dns-57d769cc4f-2lkc7\" (UID: \"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.034723 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-2lkc7\" (UID: \"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.034851 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-config\") pod \"dnsmasq-dns-57d769cc4f-2lkc7\" (UID: \"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.035974 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-config\") pod \"dnsmasq-dns-57d769cc4f-2lkc7\" (UID: \"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.036786 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-2lkc7\" (UID: \"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.062083 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq64l\" (UniqueName: \"kubernetes.io/projected/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-kube-api-access-jq64l\") pod \"dnsmasq-dns-57d769cc4f-2lkc7\" (UID: \"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.080190 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.462452 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.463868 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.468634 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.468862 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.470124 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.470236 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.470447 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.470793 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.470940 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xdhxg" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.523588 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.579455 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-44kpk"] Dec 03 11:10:46 crc kubenswrapper[4646]: W1203 11:10:46.583251 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod842907e9_4a8c_494c_a7e1_1d0686eaeb7f.slice/crio-821f5fa424d4a2956faa02e29d5e54ea5799b289c74adfb6bdd0ec18551c6bdf WatchSource:0}: Error finding container 821f5fa424d4a2956faa02e29d5e54ea5799b289c74adfb6bdd0ec18551c6bdf: Status 404 returned error can't find the container with id 821f5fa424d4a2956faa02e29d5e54ea5799b289c74adfb6bdd0ec18551c6bdf Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.642025 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.642184 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.642367 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.642457 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.642517 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.642577 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/26f0a162-8af0-438b-b48e-5bb76cb9a645-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.642615 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/26f0a162-8af0-438b-b48e-5bb76cb9a645-pod-info\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.642653 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-config-data\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.642684 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.642709 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-server-conf\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.642766 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gqgx\" (UniqueName: \"kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-kube-api-access-5gqgx\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: W1203 11:10:46.678715 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod418bf4f2_e18f_469b_ac6b_6fcb8d0ef6e0.slice/crio-febb4d1775932665c5a9d72d3386a9127e1a9c0a10e2222b4d60fbfe366b9f3c WatchSource:0}: Error finding container febb4d1775932665c5a9d72d3386a9127e1a9c0a10e2222b4d60fbfe366b9f3c: Status 404 returned error can't find the container with id febb4d1775932665c5a9d72d3386a9127e1a9c0a10e2222b4d60fbfe366b9f3c Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.679663 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2lkc7"] Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.716039 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" event={"ID":"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0","Type":"ContainerStarted","Data":"febb4d1775932665c5a9d72d3386a9127e1a9c0a10e2222b4d60fbfe366b9f3c"} Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.718855 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-44kpk" event={"ID":"842907e9-4a8c-494c-a7e1-1d0686eaeb7f","Type":"ContainerStarted","Data":"821f5fa424d4a2956faa02e29d5e54ea5799b289c74adfb6bdd0ec18551c6bdf"} Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.744595 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.744642 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.744666 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/26f0a162-8af0-438b-b48e-5bb76cb9a645-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.744686 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/26f0a162-8af0-438b-b48e-5bb76cb9a645-pod-info\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.744706 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-config-data\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.744722 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.744737 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-server-conf\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.744770 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gqgx\" (UniqueName: \"kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-kube-api-access-5gqgx\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.744794 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.744828 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.744863 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.745205 4646 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.757046 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-server-conf\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.758796 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/26f0a162-8af0-438b-b48e-5bb76cb9a645-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.764636 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.765903 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.766490 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-config-data\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.770711 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.771599 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.774988 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/26f0a162-8af0-438b-b48e-5bb76cb9a645-pod-info\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.785434 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gqgx\" (UniqueName: \"kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-kube-api-access-5gqgx\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.791724 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.801887 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.804682 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.886678 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.888263 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.898165 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-c4pm6" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.898315 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.898498 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.898598 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.898689 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.901968 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.902074 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 11:10:46 crc kubenswrapper[4646]: I1203 11:10:46.911488 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.055649 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.055923 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a6ab009e-7121-43c2-9471-ea5145e118d6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.055961 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a6ab009e-7121-43c2-9471-ea5145e118d6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.056011 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.056135 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.056191 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.056262 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.056293 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.056371 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggj5c\" (UniqueName: \"kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-kube-api-access-ggj5c\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.056472 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.056496 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.162563 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.162610 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a6ab009e-7121-43c2-9471-ea5145e118d6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.162637 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a6ab009e-7121-43c2-9471-ea5145e118d6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.163361 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.163400 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.163422 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.163449 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.163470 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.163497 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggj5c\" (UniqueName: \"kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-kube-api-access-ggj5c\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.163547 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.163563 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.163675 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.164638 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.164781 4646 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.169544 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.169543 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.170571 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a6ab009e-7121-43c2-9471-ea5145e118d6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.170771 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.170956 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.181095 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a6ab009e-7121-43c2-9471-ea5145e118d6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.192995 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggj5c\" (UniqueName: \"kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-kube-api-access-ggj5c\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.196221 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.203268 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.302231 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.426806 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 11:10:47 crc kubenswrapper[4646]: W1203 11:10:47.505414 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26f0a162_8af0_438b_b48e_5bb76cb9a645.slice/crio-f8b457d8b1350ba31368c1d3d533fc52d276eb34641e7e28080ea22454b88056 WatchSource:0}: Error finding container f8b457d8b1350ba31368c1d3d533fc52d276eb34641e7e28080ea22454b88056: Status 404 returned error can't find the container with id f8b457d8b1350ba31368c1d3d533fc52d276eb34641e7e28080ea22454b88056 Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.736322 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"26f0a162-8af0-438b-b48e-5bb76cb9a645","Type":"ContainerStarted","Data":"f8b457d8b1350ba31368c1d3d533fc52d276eb34641e7e28080ea22454b88056"} Dec 03 11:10:47 crc kubenswrapper[4646]: I1203 11:10:47.874922 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.102292 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.103878 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.110408 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.111160 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.111278 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-2bklt" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.112346 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.116991 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.133993 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.200357 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.200447 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d29b2323-9322-47f0-983c-bba7a541f4e5-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.200485 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d29b2323-9322-47f0-983c-bba7a541f4e5-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.200514 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29b2323-9322-47f0-983c-bba7a541f4e5-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.200544 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d29b2323-9322-47f0-983c-bba7a541f4e5-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.200646 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d29b2323-9322-47f0-983c-bba7a541f4e5-config-data-default\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.200909 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d29b2323-9322-47f0-983c-bba7a541f4e5-kolla-config\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.200936 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drgtx\" (UniqueName: \"kubernetes.io/projected/d29b2323-9322-47f0-983c-bba7a541f4e5-kube-api-access-drgtx\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.314446 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.314530 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d29b2323-9322-47f0-983c-bba7a541f4e5-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.314558 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d29b2323-9322-47f0-983c-bba7a541f4e5-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.314601 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29b2323-9322-47f0-983c-bba7a541f4e5-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.314635 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d29b2323-9322-47f0-983c-bba7a541f4e5-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.314666 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d29b2323-9322-47f0-983c-bba7a541f4e5-config-data-default\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.314808 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d29b2323-9322-47f0-983c-bba7a541f4e5-kolla-config\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.314834 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drgtx\" (UniqueName: \"kubernetes.io/projected/d29b2323-9322-47f0-983c-bba7a541f4e5-kube-api-access-drgtx\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.315539 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d29b2323-9322-47f0-983c-bba7a541f4e5-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.317664 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d29b2323-9322-47f0-983c-bba7a541f4e5-config-data-default\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.318792 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d29b2323-9322-47f0-983c-bba7a541f4e5-kolla-config\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.319095 4646 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.330558 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29b2323-9322-47f0-983c-bba7a541f4e5-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.331838 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d29b2323-9322-47f0-983c-bba7a541f4e5-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.340003 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d29b2323-9322-47f0-983c-bba7a541f4e5-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.364510 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drgtx\" (UniqueName: \"kubernetes.io/projected/d29b2323-9322-47f0-983c-bba7a541f4e5-kube-api-access-drgtx\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.446798 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"d29b2323-9322-47f0-983c-bba7a541f4e5\") " pod="openstack/openstack-galera-0" Dec 03 11:10:48 crc kubenswrapper[4646]: I1203 11:10:48.460806 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.419360 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.427537 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.430083 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.430584 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.430817 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.431056 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.431177 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-btc7b" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.539454 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0497fe5e-a52c-4f64-9517-e10eb809829a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.539538 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0497fe5e-a52c-4f64-9517-e10eb809829a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.539575 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0497fe5e-a52c-4f64-9517-e10eb809829a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.539593 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0497fe5e-a52c-4f64-9517-e10eb809829a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.539828 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0497fe5e-a52c-4f64-9517-e10eb809829a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.539983 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.540301 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0497fe5e-a52c-4f64-9517-e10eb809829a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.540351 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsmnc\" (UniqueName: \"kubernetes.io/projected/0497fe5e-a52c-4f64-9517-e10eb809829a-kube-api-access-rsmnc\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.643900 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0497fe5e-a52c-4f64-9517-e10eb809829a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.643955 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsmnc\" (UniqueName: \"kubernetes.io/projected/0497fe5e-a52c-4f64-9517-e10eb809829a-kube-api-access-rsmnc\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.643992 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0497fe5e-a52c-4f64-9517-e10eb809829a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.644050 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0497fe5e-a52c-4f64-9517-e10eb809829a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.644079 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0497fe5e-a52c-4f64-9517-e10eb809829a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.644100 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0497fe5e-a52c-4f64-9517-e10eb809829a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.644153 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0497fe5e-a52c-4f64-9517-e10eb809829a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.644203 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.645696 4646 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.646399 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0497fe5e-a52c-4f64-9517-e10eb809829a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.646786 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0497fe5e-a52c-4f64-9517-e10eb809829a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.646833 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0497fe5e-a52c-4f64-9517-e10eb809829a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.648012 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0497fe5e-a52c-4f64-9517-e10eb809829a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.653415 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0497fe5e-a52c-4f64-9517-e10eb809829a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.665739 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsmnc\" (UniqueName: \"kubernetes.io/projected/0497fe5e-a52c-4f64-9517-e10eb809829a-kube-api-access-rsmnc\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.666858 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0497fe5e-a52c-4f64-9517-e10eb809829a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.703487 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0497fe5e-a52c-4f64-9517-e10eb809829a\") " pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.742542 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.744022 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.749626 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-cwslz" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.749710 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.749825 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.759837 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.791381 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.864707 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/493009da-44cc-43ec-8164-4c1c20ab94e1-config-data\") pod \"memcached-0\" (UID: \"493009da-44cc-43ec-8164-4c1c20ab94e1\") " pod="openstack/memcached-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.864777 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493009da-44cc-43ec-8164-4c1c20ab94e1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"493009da-44cc-43ec-8164-4c1c20ab94e1\") " pod="openstack/memcached-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.864896 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/493009da-44cc-43ec-8164-4c1c20ab94e1-kolla-config\") pod \"memcached-0\" (UID: \"493009da-44cc-43ec-8164-4c1c20ab94e1\") " pod="openstack/memcached-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.864943 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/493009da-44cc-43ec-8164-4c1c20ab94e1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"493009da-44cc-43ec-8164-4c1c20ab94e1\") " pod="openstack/memcached-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.864985 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbvkh\" (UniqueName: \"kubernetes.io/projected/493009da-44cc-43ec-8164-4c1c20ab94e1-kube-api-access-sbvkh\") pod \"memcached-0\" (UID: \"493009da-44cc-43ec-8164-4c1c20ab94e1\") " pod="openstack/memcached-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.966165 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/493009da-44cc-43ec-8164-4c1c20ab94e1-config-data\") pod \"memcached-0\" (UID: \"493009da-44cc-43ec-8164-4c1c20ab94e1\") " pod="openstack/memcached-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.966564 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493009da-44cc-43ec-8164-4c1c20ab94e1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"493009da-44cc-43ec-8164-4c1c20ab94e1\") " pod="openstack/memcached-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.967866 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/493009da-44cc-43ec-8164-4c1c20ab94e1-config-data\") pod \"memcached-0\" (UID: \"493009da-44cc-43ec-8164-4c1c20ab94e1\") " pod="openstack/memcached-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.967995 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/493009da-44cc-43ec-8164-4c1c20ab94e1-kolla-config\") pod \"memcached-0\" (UID: \"493009da-44cc-43ec-8164-4c1c20ab94e1\") " pod="openstack/memcached-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.967393 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/493009da-44cc-43ec-8164-4c1c20ab94e1-kolla-config\") pod \"memcached-0\" (UID: \"493009da-44cc-43ec-8164-4c1c20ab94e1\") " pod="openstack/memcached-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.968417 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/493009da-44cc-43ec-8164-4c1c20ab94e1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"493009da-44cc-43ec-8164-4c1c20ab94e1\") " pod="openstack/memcached-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.968460 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbvkh\" (UniqueName: \"kubernetes.io/projected/493009da-44cc-43ec-8164-4c1c20ab94e1-kube-api-access-sbvkh\") pod \"memcached-0\" (UID: \"493009da-44cc-43ec-8164-4c1c20ab94e1\") " pod="openstack/memcached-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.974488 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/493009da-44cc-43ec-8164-4c1c20ab94e1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"493009da-44cc-43ec-8164-4c1c20ab94e1\") " pod="openstack/memcached-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.987137 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbvkh\" (UniqueName: \"kubernetes.io/projected/493009da-44cc-43ec-8164-4c1c20ab94e1-kube-api-access-sbvkh\") pod \"memcached-0\" (UID: \"493009da-44cc-43ec-8164-4c1c20ab94e1\") " pod="openstack/memcached-0" Dec 03 11:10:49 crc kubenswrapper[4646]: I1203 11:10:49.989238 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493009da-44cc-43ec-8164-4c1c20ab94e1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"493009da-44cc-43ec-8164-4c1c20ab94e1\") " pod="openstack/memcached-0" Dec 03 11:10:50 crc kubenswrapper[4646]: I1203 11:10:50.076059 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 11:10:51 crc kubenswrapper[4646]: I1203 11:10:51.933740 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 11:10:51 crc kubenswrapper[4646]: I1203 11:10:51.938568 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 11:10:51 crc kubenswrapper[4646]: I1203 11:10:51.944230 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-5sx5t" Dec 03 11:10:52 crc kubenswrapper[4646]: I1203 11:10:52.003230 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqgdl\" (UniqueName: \"kubernetes.io/projected/86ef1c10-7d50-4089-a463-676380821fe3-kube-api-access-nqgdl\") pod \"kube-state-metrics-0\" (UID: \"86ef1c10-7d50-4089-a463-676380821fe3\") " pod="openstack/kube-state-metrics-0" Dec 03 11:10:52 crc kubenswrapper[4646]: I1203 11:10:52.006071 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 11:10:52 crc kubenswrapper[4646]: I1203 11:10:52.104291 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqgdl\" (UniqueName: \"kubernetes.io/projected/86ef1c10-7d50-4089-a463-676380821fe3-kube-api-access-nqgdl\") pod \"kube-state-metrics-0\" (UID: \"86ef1c10-7d50-4089-a463-676380821fe3\") " pod="openstack/kube-state-metrics-0" Dec 03 11:10:52 crc kubenswrapper[4646]: I1203 11:10:52.125552 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqgdl\" (UniqueName: \"kubernetes.io/projected/86ef1c10-7d50-4089-a463-676380821fe3-kube-api-access-nqgdl\") pod \"kube-state-metrics-0\" (UID: \"86ef1c10-7d50-4089-a463-676380821fe3\") " pod="openstack/kube-state-metrics-0" Dec 03 11:10:52 crc kubenswrapper[4646]: I1203 11:10:52.277680 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.357155 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rjmpb"] Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.358881 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.372091 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rjmpb"] Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.372974 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.373217 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-d48d4" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.374082 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.377946 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-5v9pv"] Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.379527 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.391959 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-5v9pv"] Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.459147 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-var-run\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.459205 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7606fd8-32a7-402e-bac9-bfe718af9019-scripts\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.459232 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-scripts\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.459280 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7606fd8-32a7-402e-bac9-bfe718af9019-combined-ca-bundle\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.459313 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f7606fd8-32a7-402e-bac9-bfe718af9019-var-log-ovn\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.459374 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw44b\" (UniqueName: \"kubernetes.io/projected/f7606fd8-32a7-402e-bac9-bfe718af9019-kube-api-access-fw44b\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.459399 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f7606fd8-32a7-402e-bac9-bfe718af9019-var-run\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.459424 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-var-log\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.459454 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbk58\" (UniqueName: \"kubernetes.io/projected/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-kube-api-access-fbk58\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.459479 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-etc-ovs\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.459500 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-var-lib\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.459516 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f7606fd8-32a7-402e-bac9-bfe718af9019-var-run-ovn\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.459542 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7606fd8-32a7-402e-bac9-bfe718af9019-ovn-controller-tls-certs\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.560450 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7606fd8-32a7-402e-bac9-bfe718af9019-combined-ca-bundle\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.560500 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f7606fd8-32a7-402e-bac9-bfe718af9019-var-log-ovn\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.560544 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw44b\" (UniqueName: \"kubernetes.io/projected/f7606fd8-32a7-402e-bac9-bfe718af9019-kube-api-access-fw44b\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.560569 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f7606fd8-32a7-402e-bac9-bfe718af9019-var-run\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.560607 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-var-log\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.560645 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbk58\" (UniqueName: \"kubernetes.io/projected/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-kube-api-access-fbk58\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.560668 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-etc-ovs\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.560685 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-var-lib\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.560700 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f7606fd8-32a7-402e-bac9-bfe718af9019-var-run-ovn\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.560763 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7606fd8-32a7-402e-bac9-bfe718af9019-ovn-controller-tls-certs\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.560782 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-var-run\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.560798 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7606fd8-32a7-402e-bac9-bfe718af9019-scripts\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.560812 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-scripts\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.561455 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f7606fd8-32a7-402e-bac9-bfe718af9019-var-log-ovn\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.561576 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-var-log\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.561611 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-var-lib\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.561683 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-var-run\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.561811 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f7606fd8-32a7-402e-bac9-bfe718af9019-var-run\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.561820 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-etc-ovs\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.561818 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f7606fd8-32a7-402e-bac9-bfe718af9019-var-run-ovn\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.563095 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-scripts\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.563765 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7606fd8-32a7-402e-bac9-bfe718af9019-scripts\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.566889 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7606fd8-32a7-402e-bac9-bfe718af9019-ovn-controller-tls-certs\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.572938 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7606fd8-32a7-402e-bac9-bfe718af9019-combined-ca-bundle\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.578125 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw44b\" (UniqueName: \"kubernetes.io/projected/f7606fd8-32a7-402e-bac9-bfe718af9019-kube-api-access-fw44b\") pod \"ovn-controller-rjmpb\" (UID: \"f7606fd8-32a7-402e-bac9-bfe718af9019\") " pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.578308 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbk58\" (UniqueName: \"kubernetes.io/projected/1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4-kube-api-access-fbk58\") pod \"ovn-controller-ovs-5v9pv\" (UID: \"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4\") " pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.734358 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rjmpb" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.745748 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.964856 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:10:55 crc kubenswrapper[4646]: I1203 11:10:55.965238 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.828027 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.831464 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.836996 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-g6vg6" Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.837119 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.839497 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.840423 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.840739 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.852717 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.918851 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2ab2f042-5e66-4aa8-9a86-6687c40a675d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.919143 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ab2f042-5e66-4aa8-9a86-6687c40a675d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.919246 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.919322 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ab2f042-5e66-4aa8-9a86-6687c40a675d-config\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.919422 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2ab2f042-5e66-4aa8-9a86-6687c40a675d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.919516 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab2f042-5e66-4aa8-9a86-6687c40a675d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.919589 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ab2f042-5e66-4aa8-9a86-6687c40a675d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:58 crc kubenswrapper[4646]: I1203 11:10:58.919659 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmnwd\" (UniqueName: \"kubernetes.io/projected/2ab2f042-5e66-4aa8-9a86-6687c40a675d-kube-api-access-rmnwd\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.027591 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ab2f042-5e66-4aa8-9a86-6687c40a675d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.027641 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.027667 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ab2f042-5e66-4aa8-9a86-6687c40a675d-config\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.027685 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2ab2f042-5e66-4aa8-9a86-6687c40a675d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.027707 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab2f042-5e66-4aa8-9a86-6687c40a675d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.027723 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ab2f042-5e66-4aa8-9a86-6687c40a675d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.027736 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmnwd\" (UniqueName: \"kubernetes.io/projected/2ab2f042-5e66-4aa8-9a86-6687c40a675d-kube-api-access-rmnwd\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.027803 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2ab2f042-5e66-4aa8-9a86-6687c40a675d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.028172 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2ab2f042-5e66-4aa8-9a86-6687c40a675d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.029508 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2ab2f042-5e66-4aa8-9a86-6687c40a675d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.029799 4646 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.034329 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ab2f042-5e66-4aa8-9a86-6687c40a675d-config\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.040228 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ab2f042-5e66-4aa8-9a86-6687c40a675d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.043903 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.044723 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ab2f042-5e66-4aa8-9a86-6687c40a675d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.046081 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.056950 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.057470 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.058325 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.059380 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-l6w2c" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.059966 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab2f042-5e66-4aa8-9a86-6687c40a675d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.062829 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmnwd\" (UniqueName: \"kubernetes.io/projected/2ab2f042-5e66-4aa8-9a86-6687c40a675d-kube-api-access-rmnwd\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.062992 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"2ab2f042-5e66-4aa8-9a86-6687c40a675d\") " pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.059443 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.129106 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.129386 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcwt9\" (UniqueName: \"kubernetes.io/projected/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-kube-api-access-vcwt9\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.129434 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.129458 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.129477 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-config\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.129505 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.129559 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.129593 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.157173 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.231426 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.231699 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.231816 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.231913 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcwt9\" (UniqueName: \"kubernetes.io/projected/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-kube-api-access-vcwt9\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.232008 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.232084 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.232161 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-config\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.232256 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.232186 4646 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.232766 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.233143 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-config\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.233430 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.235830 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.235981 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.236915 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.260133 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcwt9\" (UniqueName: \"kubernetes.io/projected/155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6-kube-api-access-vcwt9\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.267743 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6\") " pod="openstack/ovsdbserver-sb-0" Dec 03 11:10:59 crc kubenswrapper[4646]: I1203 11:10:59.415775 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 11:11:01 crc kubenswrapper[4646]: I1203 11:11:01.899580 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a6ab009e-7121-43c2-9471-ea5145e118d6","Type":"ContainerStarted","Data":"fa3f6e5219fc49110caa3898635ec4817eb17ab14902cde64a0ec5ff367058ef"} Dec 03 11:11:10 crc kubenswrapper[4646]: I1203 11:11:10.000999 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rjmpb"] Dec 03 11:11:17 crc kubenswrapper[4646]: E1203 11:11:17.946043 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 11:11:17 crc kubenswrapper[4646]: E1203 11:11:17.946670 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l9ptf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-v4gdc_openstack(8cc7fc28-9dc3-4563-a188-542b67825477): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:11:17 crc kubenswrapper[4646]: E1203 11:11:17.948062 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-v4gdc" podUID="8cc7fc28-9dc3-4563-a188-542b67825477" Dec 03 11:11:18 crc kubenswrapper[4646]: E1203 11:11:18.615242 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 03 11:11:18 crc kubenswrapper[4646]: E1203 11:11:18.615912 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5gqgx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(26f0a162-8af0-438b-b48e-5bb76cb9a645): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:11:18 crc kubenswrapper[4646]: E1203 11:11:18.617165 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="26f0a162-8af0-438b-b48e-5bb76cb9a645" Dec 03 11:11:18 crc kubenswrapper[4646]: I1203 11:11:18.811732 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-v4gdc" Dec 03 11:11:18 crc kubenswrapper[4646]: I1203 11:11:18.884412 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9ptf\" (UniqueName: \"kubernetes.io/projected/8cc7fc28-9dc3-4563-a188-542b67825477-kube-api-access-l9ptf\") pod \"8cc7fc28-9dc3-4563-a188-542b67825477\" (UID: \"8cc7fc28-9dc3-4563-a188-542b67825477\") " Dec 03 11:11:18 crc kubenswrapper[4646]: I1203 11:11:18.884489 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cc7fc28-9dc3-4563-a188-542b67825477-config\") pod \"8cc7fc28-9dc3-4563-a188-542b67825477\" (UID: \"8cc7fc28-9dc3-4563-a188-542b67825477\") " Dec 03 11:11:18 crc kubenswrapper[4646]: I1203 11:11:18.885284 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cc7fc28-9dc3-4563-a188-542b67825477-config" (OuterVolumeSpecName: "config") pod "8cc7fc28-9dc3-4563-a188-542b67825477" (UID: "8cc7fc28-9dc3-4563-a188-542b67825477"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:11:18 crc kubenswrapper[4646]: I1203 11:11:18.907200 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cc7fc28-9dc3-4563-a188-542b67825477-kube-api-access-l9ptf" (OuterVolumeSpecName: "kube-api-access-l9ptf") pod "8cc7fc28-9dc3-4563-a188-542b67825477" (UID: "8cc7fc28-9dc3-4563-a188-542b67825477"). InnerVolumeSpecName "kube-api-access-l9ptf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:11:18 crc kubenswrapper[4646]: I1203 11:11:18.989826 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9ptf\" (UniqueName: \"kubernetes.io/projected/8cc7fc28-9dc3-4563-a188-542b67825477-kube-api-access-l9ptf\") on node \"crc\" DevicePath \"\"" Dec 03 11:11:18 crc kubenswrapper[4646]: I1203 11:11:18.989852 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cc7fc28-9dc3-4563-a188-542b67825477-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:11:19 crc kubenswrapper[4646]: E1203 11:11:19.031962 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 11:11:19 crc kubenswrapper[4646]: E1203 11:11:19.032116 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vlnx4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-44kpk_openstack(842907e9-4a8c-494c-a7e1-1d0686eaeb7f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:11:19 crc kubenswrapper[4646]: E1203 11:11:19.033458 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-44kpk" podUID="842907e9-4a8c-494c-a7e1-1d0686eaeb7f" Dec 03 11:11:19 crc kubenswrapper[4646]: I1203 11:11:19.033727 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-v4gdc" event={"ID":"8cc7fc28-9dc3-4563-a188-542b67825477","Type":"ContainerDied","Data":"4dddba3416bad17d1fa48e899d2bec16902768498070123fcb349bdb2b6ea31e"} Dec 03 11:11:19 crc kubenswrapper[4646]: I1203 11:11:19.033836 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-v4gdc" Dec 03 11:11:19 crc kubenswrapper[4646]: I1203 11:11:19.041402 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rjmpb" event={"ID":"f7606fd8-32a7-402e-bac9-bfe718af9019","Type":"ContainerStarted","Data":"fb9a656a843ef254e323e9d2af83eb767f917769e7d36e7b605699d96c706ad8"} Dec 03 11:11:19 crc kubenswrapper[4646]: E1203 11:11:19.046594 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="26f0a162-8af0-438b-b48e-5bb76cb9a645" Dec 03 11:11:19 crc kubenswrapper[4646]: I1203 11:11:19.113451 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-v4gdc"] Dec 03 11:11:19 crc kubenswrapper[4646]: I1203 11:11:19.118851 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-v4gdc"] Dec 03 11:11:19 crc kubenswrapper[4646]: I1203 11:11:19.324521 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 11:11:19 crc kubenswrapper[4646]: I1203 11:11:19.348170 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 11:11:19 crc kubenswrapper[4646]: E1203 11:11:19.360686 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 11:11:19 crc kubenswrapper[4646]: E1203 11:11:19.360837 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jq64l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-2lkc7_openstack(418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:11:19 crc kubenswrapper[4646]: E1203 11:11:19.362157 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" podUID="418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0" Dec 03 11:11:19 crc kubenswrapper[4646]: W1203 11:11:19.366891 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0497fe5e_a52c_4f64_9517_e10eb809829a.slice/crio-158906da9ccd6be261bdc1594d922bf1042525ae39c7731a3bddf7948831f824 WatchSource:0}: Error finding container 158906da9ccd6be261bdc1594d922bf1042525ae39c7731a3bddf7948831f824: Status 404 returned error can't find the container with id 158906da9ccd6be261bdc1594d922bf1042525ae39c7731a3bddf7948831f824 Dec 03 11:11:19 crc kubenswrapper[4646]: I1203 11:11:19.367518 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 11:11:19 crc kubenswrapper[4646]: E1203 11:11:19.368849 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 11:11:19 crc kubenswrapper[4646]: E1203 11:11:19.369027 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t945f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-zxtkc_openstack(3b7835bf-ecbc-457c-bf60-86652b766d4e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:11:19 crc kubenswrapper[4646]: E1203 11:11:19.370187 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" podUID="3b7835bf-ecbc-457c-bf60-86652b766d4e" Dec 03 11:11:19 crc kubenswrapper[4646]: W1203 11:11:19.372632 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd29b2323_9322_47f0_983c_bba7a541f4e5.slice/crio-7f46635cf6baf883bb5725799dcec139383d6fb29bd82856aeaa34d749c415f9 WatchSource:0}: Error finding container 7f46635cf6baf883bb5725799dcec139383d6fb29bd82856aeaa34d749c415f9: Status 404 returned error can't find the container with id 7f46635cf6baf883bb5725799dcec139383d6fb29bd82856aeaa34d749c415f9 Dec 03 11:11:19 crc kubenswrapper[4646]: I1203 11:11:19.591428 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 11:11:19 crc kubenswrapper[4646]: W1203 11:11:19.591805 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86ef1c10_7d50_4089_a463_676380821fe3.slice/crio-72c6447595701e96f7e326fffe009010a24f51e5bd989ca3ec26b2fc10b01a01 WatchSource:0}: Error finding container 72c6447595701e96f7e326fffe009010a24f51e5bd989ca3ec26b2fc10b01a01: Status 404 returned error can't find the container with id 72c6447595701e96f7e326fffe009010a24f51e5bd989ca3ec26b2fc10b01a01 Dec 03 11:11:19 crc kubenswrapper[4646]: I1203 11:11:19.868949 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cc7fc28-9dc3-4563-a188-542b67825477" path="/var/lib/kubelet/pods/8cc7fc28-9dc3-4563-a188-542b67825477/volumes" Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.062947 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"493009da-44cc-43ec-8164-4c1c20ab94e1","Type":"ContainerStarted","Data":"b1a9772d8ebc8f339b42a9c311b19e6496f6c628cdc79ae698c458d225b4d4cf"} Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.068590 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0497fe5e-a52c-4f64-9517-e10eb809829a","Type":"ContainerStarted","Data":"158906da9ccd6be261bdc1594d922bf1042525ae39c7731a3bddf7948831f824"} Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.074599 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"86ef1c10-7d50-4089-a463-676380821fe3","Type":"ContainerStarted","Data":"72c6447595701e96f7e326fffe009010a24f51e5bd989ca3ec26b2fc10b01a01"} Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.077327 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d29b2323-9322-47f0-983c-bba7a541f4e5","Type":"ContainerStarted","Data":"7f46635cf6baf883bb5725799dcec139383d6fb29bd82856aeaa34d749c415f9"} Dec 03 11:11:20 crc kubenswrapper[4646]: E1203 11:11:20.089014 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-44kpk" podUID="842907e9-4a8c-494c-a7e1-1d0686eaeb7f" Dec 03 11:11:20 crc kubenswrapper[4646]: E1203 11:11:20.148223 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" podUID="418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0" Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.204937 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.575940 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.733226 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b7835bf-ecbc-457c-bf60-86652b766d4e-config\") pod \"3b7835bf-ecbc-457c-bf60-86652b766d4e\" (UID: \"3b7835bf-ecbc-457c-bf60-86652b766d4e\") " Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.733280 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t945f\" (UniqueName: \"kubernetes.io/projected/3b7835bf-ecbc-457c-bf60-86652b766d4e-kube-api-access-t945f\") pod \"3b7835bf-ecbc-457c-bf60-86652b766d4e\" (UID: \"3b7835bf-ecbc-457c-bf60-86652b766d4e\") " Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.733442 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b7835bf-ecbc-457c-bf60-86652b766d4e-dns-svc\") pod \"3b7835bf-ecbc-457c-bf60-86652b766d4e\" (UID: \"3b7835bf-ecbc-457c-bf60-86652b766d4e\") " Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.733836 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b7835bf-ecbc-457c-bf60-86652b766d4e-config" (OuterVolumeSpecName: "config") pod "3b7835bf-ecbc-457c-bf60-86652b766d4e" (UID: "3b7835bf-ecbc-457c-bf60-86652b766d4e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.733854 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b7835bf-ecbc-457c-bf60-86652b766d4e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3b7835bf-ecbc-457c-bf60-86652b766d4e" (UID: "3b7835bf-ecbc-457c-bf60-86652b766d4e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.739776 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b7835bf-ecbc-457c-bf60-86652b766d4e-kube-api-access-t945f" (OuterVolumeSpecName: "kube-api-access-t945f") pod "3b7835bf-ecbc-457c-bf60-86652b766d4e" (UID: "3b7835bf-ecbc-457c-bf60-86652b766d4e"). InnerVolumeSpecName "kube-api-access-t945f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.797230 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-5v9pv"] Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.834926 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b7835bf-ecbc-457c-bf60-86652b766d4e-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.834957 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t945f\" (UniqueName: \"kubernetes.io/projected/3b7835bf-ecbc-457c-bf60-86652b766d4e-kube-api-access-t945f\") on node \"crc\" DevicePath \"\"" Dec 03 11:11:20 crc kubenswrapper[4646]: I1203 11:11:20.834966 4646 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b7835bf-ecbc-457c-bf60-86652b766d4e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 11:11:21 crc kubenswrapper[4646]: I1203 11:11:21.085972 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6","Type":"ContainerStarted","Data":"4d63a3858acd65d1485a5fbf7134af4ae285b2c267aa32ac8f2eaab29d820ce0"} Dec 03 11:11:21 crc kubenswrapper[4646]: I1203 11:11:21.087142 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" event={"ID":"3b7835bf-ecbc-457c-bf60-86652b766d4e","Type":"ContainerDied","Data":"8608997c794865c6c3ec08e66de9c5bd7450af318f91c1d0fa2938153b0fe421"} Dec 03 11:11:21 crc kubenswrapper[4646]: I1203 11:11:21.087165 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zxtkc" Dec 03 11:11:21 crc kubenswrapper[4646]: I1203 11:11:21.102785 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5v9pv" event={"ID":"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4","Type":"ContainerStarted","Data":"b86e5d88c245ed98ad9bbf209574f317020822906ca898da00e122982fa6baad"} Dec 03 11:11:21 crc kubenswrapper[4646]: I1203 11:11:21.166294 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zxtkc"] Dec 03 11:11:21 crc kubenswrapper[4646]: I1203 11:11:21.177598 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zxtkc"] Dec 03 11:11:21 crc kubenswrapper[4646]: I1203 11:11:21.386642 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 11:11:21 crc kubenswrapper[4646]: I1203 11:11:21.857305 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b7835bf-ecbc-457c-bf60-86652b766d4e" path="/var/lib/kubelet/pods/3b7835bf-ecbc-457c-bf60-86652b766d4e/volumes" Dec 03 11:11:22 crc kubenswrapper[4646]: I1203 11:11:22.122319 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2ab2f042-5e66-4aa8-9a86-6687c40a675d","Type":"ContainerStarted","Data":"5abaedaf731403b05520924dce2326bdb7b4a1e27df9529340e005831260e1f0"} Dec 03 11:11:23 crc kubenswrapper[4646]: I1203 11:11:23.133898 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a6ab009e-7121-43c2-9471-ea5145e118d6","Type":"ContainerStarted","Data":"c5a483953bdf195796f57d4f668b4cea83e28595a29fc5705a6029cc8dbf8bf3"} Dec 03 11:11:25 crc kubenswrapper[4646]: I1203 11:11:25.964888 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:11:25 crc kubenswrapper[4646]: I1203 11:11:25.965311 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:11:25 crc kubenswrapper[4646]: I1203 11:11:25.965407 4646 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 11:11:25 crc kubenswrapper[4646]: I1203 11:11:25.966295 4646 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"14e17deb53ff7184afa03eb1fff9074fe416c1136352cde34645028a19901f16"} pod="openshift-machine-config-operator/machine-config-daemon-pndzt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 11:11:25 crc kubenswrapper[4646]: I1203 11:11:25.966451 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" containerID="cri-o://14e17deb53ff7184afa03eb1fff9074fe416c1136352cde34645028a19901f16" gracePeriod=600 Dec 03 11:11:27 crc kubenswrapper[4646]: I1203 11:11:27.179452 4646 generic.go:334] "Generic (PLEG): container finished" podID="9695d947-fc43-4625-86af-a1a3e177bb27" containerID="14e17deb53ff7184afa03eb1fff9074fe416c1136352cde34645028a19901f16" exitCode=0 Dec 03 11:11:27 crc kubenswrapper[4646]: I1203 11:11:27.179529 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerDied","Data":"14e17deb53ff7184afa03eb1fff9074fe416c1136352cde34645028a19901f16"} Dec 03 11:11:27 crc kubenswrapper[4646]: I1203 11:11:27.180170 4646 scope.go:117] "RemoveContainer" containerID="41b509e57c2eee648a101eefda70c1101098be6f2f4ca7c3a458c795b0cc5f55" Dec 03 11:11:31 crc kubenswrapper[4646]: I1203 11:11:31.225551 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerStarted","Data":"dfd3243f7f8d357efda5aacf86cad097d4462b47d7d216b50c99ea94ea34af46"} Dec 03 11:11:32 crc kubenswrapper[4646]: I1203 11:11:32.235962 4646 generic.go:334] "Generic (PLEG): container finished" podID="1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4" containerID="cec08b98935a384014cdd1a669e5d73bfafda7737bcd3535734c53cc7fef55be" exitCode=0 Dec 03 11:11:32 crc kubenswrapper[4646]: I1203 11:11:32.236005 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5v9pv" event={"ID":"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4","Type":"ContainerDied","Data":"cec08b98935a384014cdd1a669e5d73bfafda7737bcd3535734c53cc7fef55be"} Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.243813 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6","Type":"ContainerStarted","Data":"1dcb45317d2131299715db67cdf3b4b54b844761c05e1a0794e365b338b6cd25"} Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.246306 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"86ef1c10-7d50-4089-a463-676380821fe3","Type":"ContainerStarted","Data":"00ca8229142fe32f0ae234fe0b2c23dc838a16d750b14c9ead9fef9b263cdb49"} Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.246468 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.247822 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rjmpb" event={"ID":"f7606fd8-32a7-402e-bac9-bfe718af9019","Type":"ContainerStarted","Data":"69ead0d30079916c1cd25f227980af9bf9fefca79af956772560277e09fbc56f"} Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.247950 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-rjmpb" Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.249182 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2ab2f042-5e66-4aa8-9a86-6687c40a675d","Type":"ContainerStarted","Data":"38a0c80fbf9f12a96d38d8690277725f67a735a62b25cb20031b692bcdb94074"} Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.257042 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5v9pv" event={"ID":"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4","Type":"ContainerStarted","Data":"f4eb94eea571d5c8c62fd15ed816ba4b45bd64d0d704dd1f1aea7bbb0641719c"} Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.257438 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.257458 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.258943 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d29b2323-9322-47f0-983c-bba7a541f4e5","Type":"ContainerStarted","Data":"356e70d7a73942af1accf51a54a15ae7981e804a1e66f199758a6b3c6667746a"} Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.262363 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"493009da-44cc-43ec-8164-4c1c20ab94e1","Type":"ContainerStarted","Data":"f7ec78b41f5f81fc5eb3e4c8e2f82447c4a6918592d6bd333a45a480f30158ab"} Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.262991 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.264502 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0497fe5e-a52c-4f64-9517-e10eb809829a","Type":"ContainerStarted","Data":"4c9370576d38760e5729b7d09fa9b785671c3f40ab5ee16aa51d8f21bbb864a8"} Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.280634 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=29.436248679 podStartE2EDuration="42.280610429s" podCreationTimestamp="2025-12-03 11:10:51 +0000 UTC" firstStartedPulling="2025-12-03 11:11:19.594475723 +0000 UTC m=+1056.057531858" lastFinishedPulling="2025-12-03 11:11:32.438837473 +0000 UTC m=+1068.901893608" observedRunningTime="2025-12-03 11:11:33.268381665 +0000 UTC m=+1069.731437790" watchObservedRunningTime="2025-12-03 11:11:33.280610429 +0000 UTC m=+1069.743666564" Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.331359 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-rjmpb" podStartSLOduration=26.381629054 podStartE2EDuration="38.331323229s" podCreationTimestamp="2025-12-03 11:10:55 +0000 UTC" firstStartedPulling="2025-12-03 11:11:18.626468137 +0000 UTC m=+1055.089524272" lastFinishedPulling="2025-12-03 11:11:30.576162312 +0000 UTC m=+1067.039218447" observedRunningTime="2025-12-03 11:11:33.328642334 +0000 UTC m=+1069.791698479" watchObservedRunningTime="2025-12-03 11:11:33.331323229 +0000 UTC m=+1069.794379364" Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.356850 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-5v9pv" podStartSLOduration=29.128414918 podStartE2EDuration="38.356835599s" podCreationTimestamp="2025-12-03 11:10:55 +0000 UTC" firstStartedPulling="2025-12-03 11:11:20.79590205 +0000 UTC m=+1057.258958185" lastFinishedPulling="2025-12-03 11:11:30.024322731 +0000 UTC m=+1066.487378866" observedRunningTime="2025-12-03 11:11:33.356757797 +0000 UTC m=+1069.819813932" watchObservedRunningTime="2025-12-03 11:11:33.356835599 +0000 UTC m=+1069.819891734" Dec 03 11:11:33 crc kubenswrapper[4646]: I1203 11:11:33.387134 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=32.850115897 podStartE2EDuration="44.387114463s" podCreationTimestamp="2025-12-03 11:10:49 +0000 UTC" firstStartedPulling="2025-12-03 11:11:19.369706355 +0000 UTC m=+1055.832762490" lastFinishedPulling="2025-12-03 11:11:30.906704921 +0000 UTC m=+1067.369761056" observedRunningTime="2025-12-03 11:11:33.384121548 +0000 UTC m=+1069.847177683" watchObservedRunningTime="2025-12-03 11:11:33.387114463 +0000 UTC m=+1069.850170598" Dec 03 11:11:34 crc kubenswrapper[4646]: I1203 11:11:34.312434 4646 generic.go:334] "Generic (PLEG): container finished" podID="418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0" containerID="8dda7455995af1dc1bae7d04451b6079c91436533ad355b2e10168a1f42a92a5" exitCode=0 Dec 03 11:11:34 crc kubenswrapper[4646]: I1203 11:11:34.312517 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" event={"ID":"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0","Type":"ContainerDied","Data":"8dda7455995af1dc1bae7d04451b6079c91436533ad355b2e10168a1f42a92a5"} Dec 03 11:11:34 crc kubenswrapper[4646]: I1203 11:11:34.316689 4646 generic.go:334] "Generic (PLEG): container finished" podID="842907e9-4a8c-494c-a7e1-1d0686eaeb7f" containerID="b7ac880e9b3145f7c9c2e3f08f594db9224f77007b0c6946ced5e4a39548ba83" exitCode=0 Dec 03 11:11:34 crc kubenswrapper[4646]: I1203 11:11:34.316766 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-44kpk" event={"ID":"842907e9-4a8c-494c-a7e1-1d0686eaeb7f","Type":"ContainerDied","Data":"b7ac880e9b3145f7c9c2e3f08f594db9224f77007b0c6946ced5e4a39548ba83"} Dec 03 11:11:34 crc kubenswrapper[4646]: I1203 11:11:34.346540 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"26f0a162-8af0-438b-b48e-5bb76cb9a645","Type":"ContainerStarted","Data":"403c513e6091509f7524e462220d0d7d652a7b5bfdbdd96c35cd55a4ea200891"} Dec 03 11:11:34 crc kubenswrapper[4646]: I1203 11:11:34.354655 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5v9pv" event={"ID":"1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4","Type":"ContainerStarted","Data":"735fc6350058a4da05fec73e1643c6b1c03b99a2243ce0b9d26465f34c567cc1"} Dec 03 11:11:36 crc kubenswrapper[4646]: I1203 11:11:36.394074 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" event={"ID":"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0","Type":"ContainerStarted","Data":"a1feb833264fc2dde0cd7c4c0f45a2e731d0459d3a1a93290d373e7db6e85b85"} Dec 03 11:11:36 crc kubenswrapper[4646]: I1203 11:11:36.394647 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" Dec 03 11:11:36 crc kubenswrapper[4646]: I1203 11:11:36.396839 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-44kpk" event={"ID":"842907e9-4a8c-494c-a7e1-1d0686eaeb7f","Type":"ContainerStarted","Data":"29a9bb14162fb19e81cd1cfba5d5b27367253a40957f73fa80de45276b6404af"} Dec 03 11:11:36 crc kubenswrapper[4646]: I1203 11:11:36.397116 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-44kpk" Dec 03 11:11:36 crc kubenswrapper[4646]: I1203 11:11:36.417202 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" podStartSLOduration=4.66870622 podStartE2EDuration="51.417187033s" podCreationTimestamp="2025-12-03 11:10:45 +0000 UTC" firstStartedPulling="2025-12-03 11:10:46.682663491 +0000 UTC m=+1023.145719626" lastFinishedPulling="2025-12-03 11:11:33.431144294 +0000 UTC m=+1069.894200439" observedRunningTime="2025-12-03 11:11:36.409181777 +0000 UTC m=+1072.872237912" watchObservedRunningTime="2025-12-03 11:11:36.417187033 +0000 UTC m=+1072.880243168" Dec 03 11:11:36 crc kubenswrapper[4646]: I1203 11:11:36.429196 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-44kpk" podStartSLOduration=4.583787247 podStartE2EDuration="51.429178531s" podCreationTimestamp="2025-12-03 11:10:45 +0000 UTC" firstStartedPulling="2025-12-03 11:10:46.588080666 +0000 UTC m=+1023.051136801" lastFinishedPulling="2025-12-03 11:11:33.43347195 +0000 UTC m=+1069.896528085" observedRunningTime="2025-12-03 11:11:36.42665139 +0000 UTC m=+1072.889707525" watchObservedRunningTime="2025-12-03 11:11:36.429178531 +0000 UTC m=+1072.892234666" Dec 03 11:11:38 crc kubenswrapper[4646]: I1203 11:11:38.416330 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2ab2f042-5e66-4aa8-9a86-6687c40a675d","Type":"ContainerStarted","Data":"b3a8efaa3102d04156d8eff2a480bae52d59003f3049bd5ac85046a794428889"} Dec 03 11:11:38 crc kubenswrapper[4646]: I1203 11:11:38.418691 4646 generic.go:334] "Generic (PLEG): container finished" podID="d29b2323-9322-47f0-983c-bba7a541f4e5" containerID="356e70d7a73942af1accf51a54a15ae7981e804a1e66f199758a6b3c6667746a" exitCode=0 Dec 03 11:11:38 crc kubenswrapper[4646]: I1203 11:11:38.418739 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d29b2323-9322-47f0-983c-bba7a541f4e5","Type":"ContainerDied","Data":"356e70d7a73942af1accf51a54a15ae7981e804a1e66f199758a6b3c6667746a"} Dec 03 11:11:38 crc kubenswrapper[4646]: I1203 11:11:38.422878 4646 generic.go:334] "Generic (PLEG): container finished" podID="0497fe5e-a52c-4f64-9517-e10eb809829a" containerID="4c9370576d38760e5729b7d09fa9b785671c3f40ab5ee16aa51d8f21bbb864a8" exitCode=0 Dec 03 11:11:38 crc kubenswrapper[4646]: I1203 11:11:38.422999 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0497fe5e-a52c-4f64-9517-e10eb809829a","Type":"ContainerDied","Data":"4c9370576d38760e5729b7d09fa9b785671c3f40ab5ee16aa51d8f21bbb864a8"} Dec 03 11:11:38 crc kubenswrapper[4646]: I1203 11:11:38.429651 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6","Type":"ContainerStarted","Data":"fa9f47f24697c34d6ece802dd7424111c907b0fe1a3e2585f82916de3b036d61"} Dec 03 11:11:38 crc kubenswrapper[4646]: I1203 11:11:38.469707 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=25.281454908 podStartE2EDuration="41.469684149s" podCreationTimestamp="2025-12-03 11:10:57 +0000 UTC" firstStartedPulling="2025-12-03 11:11:21.387275315 +0000 UTC m=+1057.850331450" lastFinishedPulling="2025-12-03 11:11:37.575504556 +0000 UTC m=+1074.038560691" observedRunningTime="2025-12-03 11:11:38.459572083 +0000 UTC m=+1074.922628218" watchObservedRunningTime="2025-12-03 11:11:38.469684149 +0000 UTC m=+1074.932740284" Dec 03 11:11:38 crc kubenswrapper[4646]: I1203 11:11:38.485459 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=24.138124839 podStartE2EDuration="41.485438783s" podCreationTimestamp="2025-12-03 11:10:57 +0000 UTC" firstStartedPulling="2025-12-03 11:11:20.237972817 +0000 UTC m=+1056.701028952" lastFinishedPulling="2025-12-03 11:11:37.585286771 +0000 UTC m=+1074.048342896" observedRunningTime="2025-12-03 11:11:38.48391036 +0000 UTC m=+1074.946966495" watchObservedRunningTime="2025-12-03 11:11:38.485438783 +0000 UTC m=+1074.948494928" Dec 03 11:11:39 crc kubenswrapper[4646]: I1203 11:11:39.157487 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 03 11:11:39 crc kubenswrapper[4646]: I1203 11:11:39.416845 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 03 11:11:39 crc kubenswrapper[4646]: I1203 11:11:39.438242 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d29b2323-9322-47f0-983c-bba7a541f4e5","Type":"ContainerStarted","Data":"6a28a8b88fdfcdf5f91c606652dbedf277af5ea03d2f92ce4c4d3fd2650894c8"} Dec 03 11:11:39 crc kubenswrapper[4646]: I1203 11:11:39.442360 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0497fe5e-a52c-4f64-9517-e10eb809829a","Type":"ContainerStarted","Data":"24c3d9a8709b45e72c5b2f42efd77f8ca19cc498216c6eccc24a592a3939b45f"} Dec 03 11:11:39 crc kubenswrapper[4646]: I1203 11:11:39.466255 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=41.81964684 podStartE2EDuration="52.466230279s" podCreationTimestamp="2025-12-03 11:10:47 +0000 UTC" firstStartedPulling="2025-12-03 11:11:19.377799093 +0000 UTC m=+1055.840855228" lastFinishedPulling="2025-12-03 11:11:30.024382532 +0000 UTC m=+1066.487438667" observedRunningTime="2025-12-03 11:11:39.464510481 +0000 UTC m=+1075.927566646" watchObservedRunningTime="2025-12-03 11:11:39.466230279 +0000 UTC m=+1075.929286404" Dec 03 11:11:39 crc kubenswrapper[4646]: I1203 11:11:39.494156 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=40.246425797 podStartE2EDuration="51.494138356s" podCreationTimestamp="2025-12-03 11:10:48 +0000 UTC" firstStartedPulling="2025-12-03 11:11:19.374119139 +0000 UTC m=+1055.837175274" lastFinishedPulling="2025-12-03 11:11:30.621831698 +0000 UTC m=+1067.084887833" observedRunningTime="2025-12-03 11:11:39.489413203 +0000 UTC m=+1075.952469348" watchObservedRunningTime="2025-12-03 11:11:39.494138356 +0000 UTC m=+1075.957194491" Dec 03 11:11:39 crc kubenswrapper[4646]: I1203 11:11:39.760378 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 03 11:11:39 crc kubenswrapper[4646]: I1203 11:11:39.760426 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 03 11:11:40 crc kubenswrapper[4646]: I1203 11:11:40.077490 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 03 11:11:40 crc kubenswrapper[4646]: I1203 11:11:40.950131 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-44kpk" Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.082530 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.135513 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-44kpk"] Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.157807 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.212942 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.416950 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.457414 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.457533 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-44kpk" podUID="842907e9-4a8c-494c-a7e1-1d0686eaeb7f" containerName="dnsmasq-dns" containerID="cri-o://29a9bb14162fb19e81cd1cfba5d5b27367253a40957f73fa80de45276b6404af" gracePeriod=10 Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.495527 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.498869 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.881831 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xltb2"] Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.883044 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.890273 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.906435 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xltb2"] Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.917962 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-xltb2\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.918389 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-xltb2\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.918564 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdxjl\" (UniqueName: \"kubernetes.io/projected/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-kube-api-access-kdxjl\") pod \"dnsmasq-dns-5bf47b49b7-xltb2\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.918699 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-config\") pod \"dnsmasq-dns-5bf47b49b7-xltb2\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.996021 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-lg5vm"] Dec 03 11:11:41 crc kubenswrapper[4646]: I1203 11:11:41.997590 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.000869 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.019653 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-xltb2\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.019896 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c6fh\" (UniqueName: \"kubernetes.io/projected/b03f673e-7c4b-4d20-b297-be13a71f8e42-kube-api-access-4c6fh\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.019982 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdxjl\" (UniqueName: \"kubernetes.io/projected/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-kube-api-access-kdxjl\") pod \"dnsmasq-dns-5bf47b49b7-xltb2\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.020055 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-config\") pod \"dnsmasq-dns-5bf47b49b7-xltb2\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.020129 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b03f673e-7c4b-4d20-b297-be13a71f8e42-ovn-rundir\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.020204 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b03f673e-7c4b-4d20-b297-be13a71f8e42-config\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.020277 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b03f673e-7c4b-4d20-b297-be13a71f8e42-ovs-rundir\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.020365 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-xltb2\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.020464 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-xltb2\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.020550 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b03f673e-7c4b-4d20-b297-be13a71f8e42-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.020630 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03f673e-7c4b-4d20-b297-be13a71f8e42-combined-ca-bundle\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.020917 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-config\") pod \"dnsmasq-dns-5bf47b49b7-xltb2\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.021301 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-xltb2\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.023743 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lg5vm"] Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.054446 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdxjl\" (UniqueName: \"kubernetes.io/projected/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-kube-api-access-kdxjl\") pod \"dnsmasq-dns-5bf47b49b7-xltb2\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.127630 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c6fh\" (UniqueName: \"kubernetes.io/projected/b03f673e-7c4b-4d20-b297-be13a71f8e42-kube-api-access-4c6fh\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.127705 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b03f673e-7c4b-4d20-b297-be13a71f8e42-ovn-rundir\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.127728 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b03f673e-7c4b-4d20-b297-be13a71f8e42-config\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.127751 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b03f673e-7c4b-4d20-b297-be13a71f8e42-ovs-rundir\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.127812 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b03f673e-7c4b-4d20-b297-be13a71f8e42-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.127831 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03f673e-7c4b-4d20-b297-be13a71f8e42-combined-ca-bundle\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.128470 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b03f673e-7c4b-4d20-b297-be13a71f8e42-ovn-rundir\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.128626 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b03f673e-7c4b-4d20-b297-be13a71f8e42-ovs-rundir\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.128723 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b03f673e-7c4b-4d20-b297-be13a71f8e42-config\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.132787 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b03f673e-7c4b-4d20-b297-be13a71f8e42-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.154838 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03f673e-7c4b-4d20-b297-be13a71f8e42-combined-ca-bundle\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.173965 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c6fh\" (UniqueName: \"kubernetes.io/projected/b03f673e-7c4b-4d20-b297-be13a71f8e42-kube-api-access-4c6fh\") pod \"ovn-controller-metrics-lg5vm\" (UID: \"b03f673e-7c4b-4d20-b297-be13a71f8e42\") " pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.188225 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xltb2"] Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.188754 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.197283 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.204190 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.232716 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.232901 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-5mtqf" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.233061 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.233643 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.276297 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.301297 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.312273 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lg5vm" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.320745 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-tqqfs"] Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.322172 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.326881 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.327045 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-tqqfs"] Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.332875 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f05884-7ce0-49e9-bf34-fdb200c78095-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.333024 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swplt\" (UniqueName: \"kubernetes.io/projected/46f05884-7ce0-49e9-bf34-fdb200c78095-kube-api-access-swplt\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.333113 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46f05884-7ce0-49e9-bf34-fdb200c78095-scripts\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.333216 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f05884-7ce0-49e9-bf34-fdb200c78095-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.333318 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/46f05884-7ce0-49e9-bf34-fdb200c78095-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.333416 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f05884-7ce0-49e9-bf34-fdb200c78095-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.333502 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f05884-7ce0-49e9-bf34-fdb200c78095-config\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.434827 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-dns-svc\") pod \"dnsmasq-dns-8554648995-tqqfs\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.434883 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f05884-7ce0-49e9-bf34-fdb200c78095-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.434901 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/46f05884-7ce0-49e9-bf34-fdb200c78095-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.434933 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-tqqfs\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.434963 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f05884-7ce0-49e9-bf34-fdb200c78095-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.434990 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f05884-7ce0-49e9-bf34-fdb200c78095-config\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.435040 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dhxv\" (UniqueName: \"kubernetes.io/projected/758d0aa0-8428-46bb-902d-f751b670c1ac-kube-api-access-7dhxv\") pod \"dnsmasq-dns-8554648995-tqqfs\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.435083 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f05884-7ce0-49e9-bf34-fdb200c78095-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.435101 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swplt\" (UniqueName: \"kubernetes.io/projected/46f05884-7ce0-49e9-bf34-fdb200c78095-kube-api-access-swplt\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.435123 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-config\") pod \"dnsmasq-dns-8554648995-tqqfs\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.435150 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46f05884-7ce0-49e9-bf34-fdb200c78095-scripts\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.435171 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-tqqfs\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.436761 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f05884-7ce0-49e9-bf34-fdb200c78095-config\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.437088 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/46f05884-7ce0-49e9-bf34-fdb200c78095-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.437649 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46f05884-7ce0-49e9-bf34-fdb200c78095-scripts\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.440849 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f05884-7ce0-49e9-bf34-fdb200c78095-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.444505 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f05884-7ce0-49e9-bf34-fdb200c78095-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.450739 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f05884-7ce0-49e9-bf34-fdb200c78095-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.458611 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swplt\" (UniqueName: \"kubernetes.io/projected/46f05884-7ce0-49e9-bf34-fdb200c78095-kube-api-access-swplt\") pod \"ovn-northd-0\" (UID: \"46f05884-7ce0-49e9-bf34-fdb200c78095\") " pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.534686 4646 generic.go:334] "Generic (PLEG): container finished" podID="842907e9-4a8c-494c-a7e1-1d0686eaeb7f" containerID="29a9bb14162fb19e81cd1cfba5d5b27367253a40957f73fa80de45276b6404af" exitCode=0 Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.534970 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-44kpk" event={"ID":"842907e9-4a8c-494c-a7e1-1d0686eaeb7f","Type":"ContainerDied","Data":"29a9bb14162fb19e81cd1cfba5d5b27367253a40957f73fa80de45276b6404af"} Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.536220 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dhxv\" (UniqueName: \"kubernetes.io/projected/758d0aa0-8428-46bb-902d-f751b670c1ac-kube-api-access-7dhxv\") pod \"dnsmasq-dns-8554648995-tqqfs\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.536300 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-config\") pod \"dnsmasq-dns-8554648995-tqqfs\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.536347 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-tqqfs\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.536391 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-dns-svc\") pod \"dnsmasq-dns-8554648995-tqqfs\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.536436 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-tqqfs\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.542739 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-tqqfs\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.543075 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-dns-svc\") pod \"dnsmasq-dns-8554648995-tqqfs\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.543829 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-tqqfs\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.551868 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-config\") pod \"dnsmasq-dns-8554648995-tqqfs\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.569353 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dhxv\" (UniqueName: \"kubernetes.io/projected/758d0aa0-8428-46bb-902d-f751b670c1ac-kube-api-access-7dhxv\") pod \"dnsmasq-dns-8554648995-tqqfs\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.586405 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.656075 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:42 crc kubenswrapper[4646]: I1203 11:11:42.915493 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xltb2"] Dec 03 11:11:43 crc kubenswrapper[4646]: I1203 11:11:43.042202 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lg5vm"] Dec 03 11:11:43 crc kubenswrapper[4646]: I1203 11:11:43.227475 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 11:11:43 crc kubenswrapper[4646]: I1203 11:11:43.287294 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-tqqfs"] Dec 03 11:11:43 crc kubenswrapper[4646]: W1203 11:11:43.306060 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod758d0aa0_8428_46bb_902d_f751b670c1ac.slice/crio-8368c35b17ca70b9803582f0023161a074fe71c1026bbdc5de27e10585b92ec0 WatchSource:0}: Error finding container 8368c35b17ca70b9803582f0023161a074fe71c1026bbdc5de27e10585b92ec0: Status 404 returned error can't find the container with id 8368c35b17ca70b9803582f0023161a074fe71c1026bbdc5de27e10585b92ec0 Dec 03 11:11:43 crc kubenswrapper[4646]: I1203 11:11:43.542778 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"46f05884-7ce0-49e9-bf34-fdb200c78095","Type":"ContainerStarted","Data":"74a3fddb6bd24e6988866fbeeba4574ad8a39d98ee1f3efd3eb74632f2bf7658"} Dec 03 11:11:43 crc kubenswrapper[4646]: I1203 11:11:43.543689 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-tqqfs" event={"ID":"758d0aa0-8428-46bb-902d-f751b670c1ac","Type":"ContainerStarted","Data":"8368c35b17ca70b9803582f0023161a074fe71c1026bbdc5de27e10585b92ec0"} Dec 03 11:11:43 crc kubenswrapper[4646]: I1203 11:11:43.544704 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" event={"ID":"438247cd-d18e-4e07-a2b8-2b6a1b73e18d","Type":"ContainerStarted","Data":"e6622d947e4bd57fface723ff94c75ea472282003414898629e4dce550d45e5a"} Dec 03 11:11:43 crc kubenswrapper[4646]: I1203 11:11:43.545620 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lg5vm" event={"ID":"b03f673e-7c4b-4d20-b297-be13a71f8e42","Type":"ContainerStarted","Data":"d9692039261a1add795b795679c894fb549673fea1be6c9cc35fea2faafeaca0"} Dec 03 11:11:45 crc kubenswrapper[4646]: I1203 11:11:45.585800 4646 generic.go:334] "Generic (PLEG): container finished" podID="758d0aa0-8428-46bb-902d-f751b670c1ac" containerID="3a09cbb4ef5c13779b35dbbb2e555bfad93b0a9b0ccb80f40cd03bc9dc1c1f2d" exitCode=0 Dec 03 11:11:45 crc kubenswrapper[4646]: I1203 11:11:45.586007 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-tqqfs" event={"ID":"758d0aa0-8428-46bb-902d-f751b670c1ac","Type":"ContainerDied","Data":"3a09cbb4ef5c13779b35dbbb2e555bfad93b0a9b0ccb80f40cd03bc9dc1c1f2d"} Dec 03 11:11:45 crc kubenswrapper[4646]: I1203 11:11:45.592685 4646 generic.go:334] "Generic (PLEG): container finished" podID="438247cd-d18e-4e07-a2b8-2b6a1b73e18d" containerID="8680051343cc7bb0cf9830d3aa55add3bc25247a73a7e7f4a4114f67623f99e9" exitCode=0 Dec 03 11:11:45 crc kubenswrapper[4646]: I1203 11:11:45.592739 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" event={"ID":"438247cd-d18e-4e07-a2b8-2b6a1b73e18d","Type":"ContainerDied","Data":"8680051343cc7bb0cf9830d3aa55add3bc25247a73a7e7f4a4114f67623f99e9"} Dec 03 11:11:45 crc kubenswrapper[4646]: I1203 11:11:45.595390 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lg5vm" event={"ID":"b03f673e-7c4b-4d20-b297-be13a71f8e42","Type":"ContainerStarted","Data":"93f936db4534d2a2171959ccc89db9085971050dd8eb43de93956ea19fd73ada"} Dec 03 11:11:45 crc kubenswrapper[4646]: I1203 11:11:45.652885 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-lg5vm" podStartSLOduration=4.652866788 podStartE2EDuration="4.652866788s" podCreationTimestamp="2025-12-03 11:11:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:11:45.646725335 +0000 UTC m=+1082.109781480" watchObservedRunningTime="2025-12-03 11:11:45.652866788 +0000 UTC m=+1082.115922913" Dec 03 11:11:45 crc kubenswrapper[4646]: I1203 11:11:45.797593 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-44kpk" Dec 03 11:11:45 crc kubenswrapper[4646]: I1203 11:11:45.868113 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlnx4\" (UniqueName: \"kubernetes.io/projected/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-kube-api-access-vlnx4\") pod \"842907e9-4a8c-494c-a7e1-1d0686eaeb7f\" (UID: \"842907e9-4a8c-494c-a7e1-1d0686eaeb7f\") " Dec 03 11:11:45 crc kubenswrapper[4646]: I1203 11:11:45.868440 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-dns-svc\") pod \"842907e9-4a8c-494c-a7e1-1d0686eaeb7f\" (UID: \"842907e9-4a8c-494c-a7e1-1d0686eaeb7f\") " Dec 03 11:11:45 crc kubenswrapper[4646]: I1203 11:11:45.868510 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-config\") pod \"842907e9-4a8c-494c-a7e1-1d0686eaeb7f\" (UID: \"842907e9-4a8c-494c-a7e1-1d0686eaeb7f\") " Dec 03 11:11:45 crc kubenswrapper[4646]: I1203 11:11:45.878035 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-kube-api-access-vlnx4" (OuterVolumeSpecName: "kube-api-access-vlnx4") pod "842907e9-4a8c-494c-a7e1-1d0686eaeb7f" (UID: "842907e9-4a8c-494c-a7e1-1d0686eaeb7f"). InnerVolumeSpecName "kube-api-access-vlnx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:11:45 crc kubenswrapper[4646]: I1203 11:11:45.975805 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlnx4\" (UniqueName: \"kubernetes.io/projected/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-kube-api-access-vlnx4\") on node \"crc\" DevicePath \"\"" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.098176 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-config" (OuterVolumeSpecName: "config") pod "842907e9-4a8c-494c-a7e1-1d0686eaeb7f" (UID: "842907e9-4a8c-494c-a7e1-1d0686eaeb7f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.107656 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "842907e9-4a8c-494c-a7e1-1d0686eaeb7f" (UID: "842907e9-4a8c-494c-a7e1-1d0686eaeb7f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.181718 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.181745 4646 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/842907e9-4a8c-494c-a7e1-1d0686eaeb7f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.187030 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.286839 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdxjl\" (UniqueName: \"kubernetes.io/projected/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-kube-api-access-kdxjl\") pod \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.287022 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-ovsdbserver-nb\") pod \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.287076 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-dns-svc\") pod \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.287129 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-config\") pod \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\" (UID: \"438247cd-d18e-4e07-a2b8-2b6a1b73e18d\") " Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.300524 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-kube-api-access-kdxjl" (OuterVolumeSpecName: "kube-api-access-kdxjl") pod "438247cd-d18e-4e07-a2b8-2b6a1b73e18d" (UID: "438247cd-d18e-4e07-a2b8-2b6a1b73e18d"). InnerVolumeSpecName "kube-api-access-kdxjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.318896 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "438247cd-d18e-4e07-a2b8-2b6a1b73e18d" (UID: "438247cd-d18e-4e07-a2b8-2b6a1b73e18d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.350515 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "438247cd-d18e-4e07-a2b8-2b6a1b73e18d" (UID: "438247cd-d18e-4e07-a2b8-2b6a1b73e18d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.379831 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-config" (OuterVolumeSpecName: "config") pod "438247cd-d18e-4e07-a2b8-2b6a1b73e18d" (UID: "438247cd-d18e-4e07-a2b8-2b6a1b73e18d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.388439 4646 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.388469 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.388479 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdxjl\" (UniqueName: \"kubernetes.io/projected/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-kube-api-access-kdxjl\") on node \"crc\" DevicePath \"\"" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.388493 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/438247cd-d18e-4e07-a2b8-2b6a1b73e18d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.612520 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-tqqfs" event={"ID":"758d0aa0-8428-46bb-902d-f751b670c1ac","Type":"ContainerStarted","Data":"8adc4d7cca6da85aa4a5b3e75931c2edb7f7d3204c30b6e47348f912816cb87f"} Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.612855 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.626115 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-44kpk" event={"ID":"842907e9-4a8c-494c-a7e1-1d0686eaeb7f","Type":"ContainerDied","Data":"821f5fa424d4a2956faa02e29d5e54ea5799b289c74adfb6bdd0ec18551c6bdf"} Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.626187 4646 scope.go:117] "RemoveContainer" containerID="29a9bb14162fb19e81cd1cfba5d5b27367253a40957f73fa80de45276b6404af" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.626353 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-44kpk" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.644398 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" event={"ID":"438247cd-d18e-4e07-a2b8-2b6a1b73e18d","Type":"ContainerDied","Data":"e6622d947e4bd57fface723ff94c75ea472282003414898629e4dce550d45e5a"} Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.644447 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-xltb2" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.711087 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-tqqfs" podStartSLOduration=4.711072618 podStartE2EDuration="4.711072618s" podCreationTimestamp="2025-12-03 11:11:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:11:46.659446162 +0000 UTC m=+1083.122502307" watchObservedRunningTime="2025-12-03 11:11:46.711072618 +0000 UTC m=+1083.174128753" Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.715392 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-44kpk"] Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.727269 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-44kpk"] Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.775782 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xltb2"] Dec 03 11:11:46 crc kubenswrapper[4646]: I1203 11:11:46.783984 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xltb2"] Dec 03 11:11:47 crc kubenswrapper[4646]: I1203 11:11:47.480162 4646 scope.go:117] "RemoveContainer" containerID="b7ac880e9b3145f7c9c2e3f08f594db9224f77007b0c6946ced5e4a39548ba83" Dec 03 11:11:47 crc kubenswrapper[4646]: I1203 11:11:47.537490 4646 scope.go:117] "RemoveContainer" containerID="8680051343cc7bb0cf9830d3aa55add3bc25247a73a7e7f4a4114f67623f99e9" Dec 03 11:11:47 crc kubenswrapper[4646]: I1203 11:11:47.859909 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="438247cd-d18e-4e07-a2b8-2b6a1b73e18d" path="/var/lib/kubelet/pods/438247cd-d18e-4e07-a2b8-2b6a1b73e18d/volumes" Dec 03 11:11:47 crc kubenswrapper[4646]: I1203 11:11:47.860694 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="842907e9-4a8c-494c-a7e1-1d0686eaeb7f" path="/var/lib/kubelet/pods/842907e9-4a8c-494c-a7e1-1d0686eaeb7f/volumes" Dec 03 11:11:47 crc kubenswrapper[4646]: I1203 11:11:47.891934 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 03 11:11:48 crc kubenswrapper[4646]: I1203 11:11:48.007531 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 03 11:11:48 crc kubenswrapper[4646]: I1203 11:11:48.461747 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 03 11:11:48 crc kubenswrapper[4646]: I1203 11:11:48.461802 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 03 11:11:48 crc kubenswrapper[4646]: I1203 11:11:48.660723 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"46f05884-7ce0-49e9-bf34-fdb200c78095","Type":"ContainerStarted","Data":"8baf8d7ee89fa0341c4d89953c5cde68e0608bb4969e6a64b22efa081be7d61c"} Dec 03 11:11:48 crc kubenswrapper[4646]: I1203 11:11:48.660772 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"46f05884-7ce0-49e9-bf34-fdb200c78095","Type":"ContainerStarted","Data":"f627e29378398d0bb2854f4810ad14fe510da73f61afb987dcf9297a631291f9"} Dec 03 11:11:48 crc kubenswrapper[4646]: I1203 11:11:48.660899 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 03 11:11:48 crc kubenswrapper[4646]: I1203 11:11:48.687716 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.408946453 podStartE2EDuration="6.687694364s" podCreationTimestamp="2025-12-03 11:11:42 +0000 UTC" firstStartedPulling="2025-12-03 11:11:43.260737126 +0000 UTC m=+1079.723793261" lastFinishedPulling="2025-12-03 11:11:47.539485037 +0000 UTC m=+1084.002541172" observedRunningTime="2025-12-03 11:11:48.67763104 +0000 UTC m=+1085.140687175" watchObservedRunningTime="2025-12-03 11:11:48.687694364 +0000 UTC m=+1085.150750499" Dec 03 11:11:50 crc kubenswrapper[4646]: I1203 11:11:50.692517 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 03 11:11:50 crc kubenswrapper[4646]: I1203 11:11:50.769875 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 03 11:11:52 crc kubenswrapper[4646]: I1203 11:11:52.659500 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:11:52 crc kubenswrapper[4646]: I1203 11:11:52.750432 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2lkc7"] Dec 03 11:11:52 crc kubenswrapper[4646]: I1203 11:11:52.750703 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" podUID="418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0" containerName="dnsmasq-dns" containerID="cri-o://a1feb833264fc2dde0cd7c4c0f45a2e731d0459d3a1a93290d373e7db6e85b85" gracePeriod=10 Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.295873 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.433095 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-dns-svc\") pod \"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0\" (UID: \"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0\") " Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.433183 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-config\") pod \"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0\" (UID: \"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0\") " Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.433372 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq64l\" (UniqueName: \"kubernetes.io/projected/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-kube-api-access-jq64l\") pod \"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0\" (UID: \"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0\") " Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.443951 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-kube-api-access-jq64l" (OuterVolumeSpecName: "kube-api-access-jq64l") pod "418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0" (UID: "418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0"). InnerVolumeSpecName "kube-api-access-jq64l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.475364 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0" (UID: "418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.476361 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-config" (OuterVolumeSpecName: "config") pod "418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0" (UID: "418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.535173 4646 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.535204 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.535213 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq64l\" (UniqueName: \"kubernetes.io/projected/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0-kube-api-access-jq64l\") on node \"crc\" DevicePath \"\"" Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.707685 4646 generic.go:334] "Generic (PLEG): container finished" podID="418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0" containerID="a1feb833264fc2dde0cd7c4c0f45a2e731d0459d3a1a93290d373e7db6e85b85" exitCode=0 Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.707749 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.707742 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" event={"ID":"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0","Type":"ContainerDied","Data":"a1feb833264fc2dde0cd7c4c0f45a2e731d0459d3a1a93290d373e7db6e85b85"} Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.707816 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2lkc7" event={"ID":"418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0","Type":"ContainerDied","Data":"febb4d1775932665c5a9d72d3386a9127e1a9c0a10e2222b4d60fbfe366b9f3c"} Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.707836 4646 scope.go:117] "RemoveContainer" containerID="a1feb833264fc2dde0cd7c4c0f45a2e731d0459d3a1a93290d373e7db6e85b85" Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.755818 4646 scope.go:117] "RemoveContainer" containerID="8dda7455995af1dc1bae7d04451b6079c91436533ad355b2e10168a1f42a92a5" Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.761974 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2lkc7"] Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.768609 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2lkc7"] Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.782595 4646 scope.go:117] "RemoveContainer" containerID="a1feb833264fc2dde0cd7c4c0f45a2e731d0459d3a1a93290d373e7db6e85b85" Dec 03 11:11:53 crc kubenswrapper[4646]: E1203 11:11:53.783087 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1feb833264fc2dde0cd7c4c0f45a2e731d0459d3a1a93290d373e7db6e85b85\": container with ID starting with a1feb833264fc2dde0cd7c4c0f45a2e731d0459d3a1a93290d373e7db6e85b85 not found: ID does not exist" containerID="a1feb833264fc2dde0cd7c4c0f45a2e731d0459d3a1a93290d373e7db6e85b85" Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.783143 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1feb833264fc2dde0cd7c4c0f45a2e731d0459d3a1a93290d373e7db6e85b85"} err="failed to get container status \"a1feb833264fc2dde0cd7c4c0f45a2e731d0459d3a1a93290d373e7db6e85b85\": rpc error: code = NotFound desc = could not find container \"a1feb833264fc2dde0cd7c4c0f45a2e731d0459d3a1a93290d373e7db6e85b85\": container with ID starting with a1feb833264fc2dde0cd7c4c0f45a2e731d0459d3a1a93290d373e7db6e85b85 not found: ID does not exist" Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.783174 4646 scope.go:117] "RemoveContainer" containerID="8dda7455995af1dc1bae7d04451b6079c91436533ad355b2e10168a1f42a92a5" Dec 03 11:11:53 crc kubenswrapper[4646]: E1203 11:11:53.783536 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dda7455995af1dc1bae7d04451b6079c91436533ad355b2e10168a1f42a92a5\": container with ID starting with 8dda7455995af1dc1bae7d04451b6079c91436533ad355b2e10168a1f42a92a5 not found: ID does not exist" containerID="8dda7455995af1dc1bae7d04451b6079c91436533ad355b2e10168a1f42a92a5" Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.783584 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dda7455995af1dc1bae7d04451b6079c91436533ad355b2e10168a1f42a92a5"} err="failed to get container status \"8dda7455995af1dc1bae7d04451b6079c91436533ad355b2e10168a1f42a92a5\": rpc error: code = NotFound desc = could not find container \"8dda7455995af1dc1bae7d04451b6079c91436533ad355b2e10168a1f42a92a5\": container with ID starting with 8dda7455995af1dc1bae7d04451b6079c91436533ad355b2e10168a1f42a92a5 not found: ID does not exist" Dec 03 11:11:53 crc kubenswrapper[4646]: I1203 11:11:53.861208 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0" path="/var/lib/kubelet/pods/418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0/volumes" Dec 03 11:11:54 crc kubenswrapper[4646]: I1203 11:11:54.716872 4646 generic.go:334] "Generic (PLEG): container finished" podID="a6ab009e-7121-43c2-9471-ea5145e118d6" containerID="c5a483953bdf195796f57d4f668b4cea83e28595a29fc5705a6029cc8dbf8bf3" exitCode=0 Dec 03 11:11:54 crc kubenswrapper[4646]: I1203 11:11:54.716936 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a6ab009e-7121-43c2-9471-ea5145e118d6","Type":"ContainerDied","Data":"c5a483953bdf195796f57d4f668b4cea83e28595a29fc5705a6029cc8dbf8bf3"} Dec 03 11:11:55 crc kubenswrapper[4646]: I1203 11:11:55.733848 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a6ab009e-7121-43c2-9471-ea5145e118d6","Type":"ContainerStarted","Data":"6a510054cfcd3acdae0fec82f5d96fad1810c4ce06e56cf7b626b54fe2b008dd"} Dec 03 11:11:55 crc kubenswrapper[4646]: I1203 11:11:55.734919 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:11:55 crc kubenswrapper[4646]: I1203 11:11:55.766767 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=52.364941225 podStartE2EDuration="1m10.766750097s" podCreationTimestamp="2025-12-03 11:10:45 +0000 UTC" firstStartedPulling="2025-12-03 11:11:00.884733428 +0000 UTC m=+1037.347789563" lastFinishedPulling="2025-12-03 11:11:19.2865423 +0000 UTC m=+1055.749598435" observedRunningTime="2025-12-03 11:11:55.756996282 +0000 UTC m=+1092.220052427" watchObservedRunningTime="2025-12-03 11:11:55.766750097 +0000 UTC m=+1092.229806222" Dec 03 11:11:57 crc kubenswrapper[4646]: I1203 11:11:57.665832 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.712366 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bdf0-account-create-update-z84rm"] Dec 03 11:11:59 crc kubenswrapper[4646]: E1203 11:11:59.713120 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0" containerName="dnsmasq-dns" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.713139 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0" containerName="dnsmasq-dns" Dec 03 11:11:59 crc kubenswrapper[4646]: E1203 11:11:59.713155 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="842907e9-4a8c-494c-a7e1-1d0686eaeb7f" containerName="init" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.713162 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="842907e9-4a8c-494c-a7e1-1d0686eaeb7f" containerName="init" Dec 03 11:11:59 crc kubenswrapper[4646]: E1203 11:11:59.713186 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="842907e9-4a8c-494c-a7e1-1d0686eaeb7f" containerName="dnsmasq-dns" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.713195 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="842907e9-4a8c-494c-a7e1-1d0686eaeb7f" containerName="dnsmasq-dns" Dec 03 11:11:59 crc kubenswrapper[4646]: E1203 11:11:59.713217 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0" containerName="init" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.713225 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0" containerName="init" Dec 03 11:11:59 crc kubenswrapper[4646]: E1203 11:11:59.713242 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="438247cd-d18e-4e07-a2b8-2b6a1b73e18d" containerName="init" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.713250 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="438247cd-d18e-4e07-a2b8-2b6a1b73e18d" containerName="init" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.713446 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="842907e9-4a8c-494c-a7e1-1d0686eaeb7f" containerName="dnsmasq-dns" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.713475 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="438247cd-d18e-4e07-a2b8-2b6a1b73e18d" containerName="init" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.713491 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="418bf4f2-e18f-469b-ac6b-6fcb8d0ef6e0" containerName="dnsmasq-dns" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.714083 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bdf0-account-create-update-z84rm" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.716540 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.730508 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bdf0-account-create-update-z84rm"] Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.818804 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-xg9bs"] Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.839915 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xg9bs" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.845242 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-xg9bs"] Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.856426 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af4fb79d-1d31-4ae4-8769-ff0e6bae317f-operator-scripts\") pod \"keystone-bdf0-account-create-update-z84rm\" (UID: \"af4fb79d-1d31-4ae4-8769-ff0e6bae317f\") " pod="openstack/keystone-bdf0-account-create-update-z84rm" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.856692 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q74z\" (UniqueName: \"kubernetes.io/projected/af4fb79d-1d31-4ae4-8769-ff0e6bae317f-kube-api-access-7q74z\") pod \"keystone-bdf0-account-create-update-z84rm\" (UID: \"af4fb79d-1d31-4ae4-8769-ff0e6bae317f\") " pod="openstack/keystone-bdf0-account-create-update-z84rm" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.958519 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af4fb79d-1d31-4ae4-8769-ff0e6bae317f-operator-scripts\") pod \"keystone-bdf0-account-create-update-z84rm\" (UID: \"af4fb79d-1d31-4ae4-8769-ff0e6bae317f\") " pod="openstack/keystone-bdf0-account-create-update-z84rm" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.958604 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcknk\" (UniqueName: \"kubernetes.io/projected/85bbeb75-a9c1-4812-916c-53bf850f1cb9-kube-api-access-pcknk\") pod \"keystone-db-create-xg9bs\" (UID: \"85bbeb75-a9c1-4812-916c-53bf850f1cb9\") " pod="openstack/keystone-db-create-xg9bs" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.958771 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85bbeb75-a9c1-4812-916c-53bf850f1cb9-operator-scripts\") pod \"keystone-db-create-xg9bs\" (UID: \"85bbeb75-a9c1-4812-916c-53bf850f1cb9\") " pod="openstack/keystone-db-create-xg9bs" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.958863 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q74z\" (UniqueName: \"kubernetes.io/projected/af4fb79d-1d31-4ae4-8769-ff0e6bae317f-kube-api-access-7q74z\") pod \"keystone-bdf0-account-create-update-z84rm\" (UID: \"af4fb79d-1d31-4ae4-8769-ff0e6bae317f\") " pod="openstack/keystone-bdf0-account-create-update-z84rm" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.959456 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af4fb79d-1d31-4ae4-8769-ff0e6bae317f-operator-scripts\") pod \"keystone-bdf0-account-create-update-z84rm\" (UID: \"af4fb79d-1d31-4ae4-8769-ff0e6bae317f\") " pod="openstack/keystone-bdf0-account-create-update-z84rm" Dec 03 11:11:59 crc kubenswrapper[4646]: I1203 11:11:59.979771 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q74z\" (UniqueName: \"kubernetes.io/projected/af4fb79d-1d31-4ae4-8769-ff0e6bae317f-kube-api-access-7q74z\") pod \"keystone-bdf0-account-create-update-z84rm\" (UID: \"af4fb79d-1d31-4ae4-8769-ff0e6bae317f\") " pod="openstack/keystone-bdf0-account-create-update-z84rm" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.004855 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-26vjv"] Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.006390 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-26vjv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.015827 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-26vjv"] Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.060873 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcknk\" (UniqueName: \"kubernetes.io/projected/85bbeb75-a9c1-4812-916c-53bf850f1cb9-kube-api-access-pcknk\") pod \"keystone-db-create-xg9bs\" (UID: \"85bbeb75-a9c1-4812-916c-53bf850f1cb9\") " pod="openstack/keystone-db-create-xg9bs" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.060939 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85bbeb75-a9c1-4812-916c-53bf850f1cb9-operator-scripts\") pod \"keystone-db-create-xg9bs\" (UID: \"85bbeb75-a9c1-4812-916c-53bf850f1cb9\") " pod="openstack/keystone-db-create-xg9bs" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.061649 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85bbeb75-a9c1-4812-916c-53bf850f1cb9-operator-scripts\") pod \"keystone-db-create-xg9bs\" (UID: \"85bbeb75-a9c1-4812-916c-53bf850f1cb9\") " pod="openstack/keystone-db-create-xg9bs" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.072275 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bdf0-account-create-update-z84rm" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.077939 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcknk\" (UniqueName: \"kubernetes.io/projected/85bbeb75-a9c1-4812-916c-53bf850f1cb9-kube-api-access-pcknk\") pod \"keystone-db-create-xg9bs\" (UID: \"85bbeb75-a9c1-4812-916c-53bf850f1cb9\") " pod="openstack/keystone-db-create-xg9bs" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.150667 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9826-account-create-update-9ldlv"] Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.151781 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9826-account-create-update-9ldlv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.155003 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.165549 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/237d9726-11ed-4d72-8135-b8ba2f14813f-operator-scripts\") pod \"placement-db-create-26vjv\" (UID: \"237d9726-11ed-4d72-8135-b8ba2f14813f\") " pod="openstack/placement-db-create-26vjv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.165686 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvjg5\" (UniqueName: \"kubernetes.io/projected/237d9726-11ed-4d72-8135-b8ba2f14813f-kube-api-access-gvjg5\") pod \"placement-db-create-26vjv\" (UID: \"237d9726-11ed-4d72-8135-b8ba2f14813f\") " pod="openstack/placement-db-create-26vjv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.169611 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9826-account-create-update-9ldlv"] Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.172045 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xg9bs" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.267210 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22952a9c-263f-40d2-8f5f-54ef029814c8-operator-scripts\") pod \"placement-9826-account-create-update-9ldlv\" (UID: \"22952a9c-263f-40d2-8f5f-54ef029814c8\") " pod="openstack/placement-9826-account-create-update-9ldlv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.267276 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvjg5\" (UniqueName: \"kubernetes.io/projected/237d9726-11ed-4d72-8135-b8ba2f14813f-kube-api-access-gvjg5\") pod \"placement-db-create-26vjv\" (UID: \"237d9726-11ed-4d72-8135-b8ba2f14813f\") " pod="openstack/placement-db-create-26vjv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.267316 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt6mf\" (UniqueName: \"kubernetes.io/projected/22952a9c-263f-40d2-8f5f-54ef029814c8-kube-api-access-xt6mf\") pod \"placement-9826-account-create-update-9ldlv\" (UID: \"22952a9c-263f-40d2-8f5f-54ef029814c8\") " pod="openstack/placement-9826-account-create-update-9ldlv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.267377 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/237d9726-11ed-4d72-8135-b8ba2f14813f-operator-scripts\") pod \"placement-db-create-26vjv\" (UID: \"237d9726-11ed-4d72-8135-b8ba2f14813f\") " pod="openstack/placement-db-create-26vjv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.272692 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/237d9726-11ed-4d72-8135-b8ba2f14813f-operator-scripts\") pod \"placement-db-create-26vjv\" (UID: \"237d9726-11ed-4d72-8135-b8ba2f14813f\") " pod="openstack/placement-db-create-26vjv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.309902 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvjg5\" (UniqueName: \"kubernetes.io/projected/237d9726-11ed-4d72-8135-b8ba2f14813f-kube-api-access-gvjg5\") pod \"placement-db-create-26vjv\" (UID: \"237d9726-11ed-4d72-8135-b8ba2f14813f\") " pod="openstack/placement-db-create-26vjv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.332862 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-26vjv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.368471 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22952a9c-263f-40d2-8f5f-54ef029814c8-operator-scripts\") pod \"placement-9826-account-create-update-9ldlv\" (UID: \"22952a9c-263f-40d2-8f5f-54ef029814c8\") " pod="openstack/placement-9826-account-create-update-9ldlv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.368551 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt6mf\" (UniqueName: \"kubernetes.io/projected/22952a9c-263f-40d2-8f5f-54ef029814c8-kube-api-access-xt6mf\") pod \"placement-9826-account-create-update-9ldlv\" (UID: \"22952a9c-263f-40d2-8f5f-54ef029814c8\") " pod="openstack/placement-9826-account-create-update-9ldlv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.369629 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22952a9c-263f-40d2-8f5f-54ef029814c8-operator-scripts\") pod \"placement-9826-account-create-update-9ldlv\" (UID: \"22952a9c-263f-40d2-8f5f-54ef029814c8\") " pod="openstack/placement-9826-account-create-update-9ldlv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.386572 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-6grfm"] Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.391696 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6grfm" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.412951 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt6mf\" (UniqueName: \"kubernetes.io/projected/22952a9c-263f-40d2-8f5f-54ef029814c8-kube-api-access-xt6mf\") pod \"placement-9826-account-create-update-9ldlv\" (UID: \"22952a9c-263f-40d2-8f5f-54ef029814c8\") " pod="openstack/placement-9826-account-create-update-9ldlv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.469061 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-11b8-account-create-update-khjqk"] Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.470207 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-11b8-account-create-update-khjqk" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.479695 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.489866 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-11b8-account-create-update-khjqk"] Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.492916 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9826-account-create-update-9ldlv" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.525343 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6grfm"] Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.573250 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjn54\" (UniqueName: \"kubernetes.io/projected/b0660b75-6739-4a9e-8cb4-ff6b15b0080f-kube-api-access-sjn54\") pod \"glance-db-create-6grfm\" (UID: \"b0660b75-6739-4a9e-8cb4-ff6b15b0080f\") " pod="openstack/glance-db-create-6grfm" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.573317 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkbht\" (UniqueName: \"kubernetes.io/projected/b601410d-7e0e-48e7-b2cb-81126c9ed96b-kube-api-access-jkbht\") pod \"glance-11b8-account-create-update-khjqk\" (UID: \"b601410d-7e0e-48e7-b2cb-81126c9ed96b\") " pod="openstack/glance-11b8-account-create-update-khjqk" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.573378 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b601410d-7e0e-48e7-b2cb-81126c9ed96b-operator-scripts\") pod \"glance-11b8-account-create-update-khjqk\" (UID: \"b601410d-7e0e-48e7-b2cb-81126c9ed96b\") " pod="openstack/glance-11b8-account-create-update-khjqk" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.573450 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0660b75-6739-4a9e-8cb4-ff6b15b0080f-operator-scripts\") pod \"glance-db-create-6grfm\" (UID: \"b0660b75-6739-4a9e-8cb4-ff6b15b0080f\") " pod="openstack/glance-db-create-6grfm" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.675395 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjn54\" (UniqueName: \"kubernetes.io/projected/b0660b75-6739-4a9e-8cb4-ff6b15b0080f-kube-api-access-sjn54\") pod \"glance-db-create-6grfm\" (UID: \"b0660b75-6739-4a9e-8cb4-ff6b15b0080f\") " pod="openstack/glance-db-create-6grfm" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.675479 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkbht\" (UniqueName: \"kubernetes.io/projected/b601410d-7e0e-48e7-b2cb-81126c9ed96b-kube-api-access-jkbht\") pod \"glance-11b8-account-create-update-khjqk\" (UID: \"b601410d-7e0e-48e7-b2cb-81126c9ed96b\") " pod="openstack/glance-11b8-account-create-update-khjqk" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.675545 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b601410d-7e0e-48e7-b2cb-81126c9ed96b-operator-scripts\") pod \"glance-11b8-account-create-update-khjqk\" (UID: \"b601410d-7e0e-48e7-b2cb-81126c9ed96b\") " pod="openstack/glance-11b8-account-create-update-khjqk" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.675629 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0660b75-6739-4a9e-8cb4-ff6b15b0080f-operator-scripts\") pod \"glance-db-create-6grfm\" (UID: \"b0660b75-6739-4a9e-8cb4-ff6b15b0080f\") " pod="openstack/glance-db-create-6grfm" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.676504 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0660b75-6739-4a9e-8cb4-ff6b15b0080f-operator-scripts\") pod \"glance-db-create-6grfm\" (UID: \"b0660b75-6739-4a9e-8cb4-ff6b15b0080f\") " pod="openstack/glance-db-create-6grfm" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.677089 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b601410d-7e0e-48e7-b2cb-81126c9ed96b-operator-scripts\") pod \"glance-11b8-account-create-update-khjqk\" (UID: \"b601410d-7e0e-48e7-b2cb-81126c9ed96b\") " pod="openstack/glance-11b8-account-create-update-khjqk" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.701175 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjn54\" (UniqueName: \"kubernetes.io/projected/b0660b75-6739-4a9e-8cb4-ff6b15b0080f-kube-api-access-sjn54\") pod \"glance-db-create-6grfm\" (UID: \"b0660b75-6739-4a9e-8cb4-ff6b15b0080f\") " pod="openstack/glance-db-create-6grfm" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.715560 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkbht\" (UniqueName: \"kubernetes.io/projected/b601410d-7e0e-48e7-b2cb-81126c9ed96b-kube-api-access-jkbht\") pod \"glance-11b8-account-create-update-khjqk\" (UID: \"b601410d-7e0e-48e7-b2cb-81126c9ed96b\") " pod="openstack/glance-11b8-account-create-update-khjqk" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.716894 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6grfm" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.798381 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-11b8-account-create-update-khjqk" Dec 03 11:12:00 crc kubenswrapper[4646]: I1203 11:12:00.808726 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bdf0-account-create-update-z84rm"] Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.086532 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-26vjv"] Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.217621 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9826-account-create-update-9ldlv"] Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.225966 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-xg9bs"] Dec 03 11:12:01 crc kubenswrapper[4646]: W1203 11:12:01.238060 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22952a9c_263f_40d2_8f5f_54ef029814c8.slice/crio-080def12d1a9da22699dc411f781315adae5d7a899878cbe1678e7873a4cf5a9 WatchSource:0}: Error finding container 080def12d1a9da22699dc411f781315adae5d7a899878cbe1678e7873a4cf5a9: Status 404 returned error can't find the container with id 080def12d1a9da22699dc411f781315adae5d7a899878cbe1678e7873a4cf5a9 Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.375661 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-11b8-account-create-update-khjqk"] Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.385441 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6grfm"] Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.787559 4646 generic.go:334] "Generic (PLEG): container finished" podID="af4fb79d-1d31-4ae4-8769-ff0e6bae317f" containerID="6d38eb32c115ee6ec2bbacc5d9d20a10414d18e0d52ce94ec15a3e4909f22f30" exitCode=0 Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.787624 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bdf0-account-create-update-z84rm" event={"ID":"af4fb79d-1d31-4ae4-8769-ff0e6bae317f","Type":"ContainerDied","Data":"6d38eb32c115ee6ec2bbacc5d9d20a10414d18e0d52ce94ec15a3e4909f22f30"} Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.787653 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bdf0-account-create-update-z84rm" event={"ID":"af4fb79d-1d31-4ae4-8769-ff0e6bae317f","Type":"ContainerStarted","Data":"d2da5ab1eca3e1964302d27e53499e0181485c15498e379c738aaa10fafe0f1c"} Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.789644 4646 generic.go:334] "Generic (PLEG): container finished" podID="b0660b75-6739-4a9e-8cb4-ff6b15b0080f" containerID="d606f21242655a18dbee2c322252b15d732ae0b5a80c672b4ea931a8386148ae" exitCode=0 Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.789938 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6grfm" event={"ID":"b0660b75-6739-4a9e-8cb4-ff6b15b0080f","Type":"ContainerDied","Data":"d606f21242655a18dbee2c322252b15d732ae0b5a80c672b4ea931a8386148ae"} Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.789963 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6grfm" event={"ID":"b0660b75-6739-4a9e-8cb4-ff6b15b0080f","Type":"ContainerStarted","Data":"935b5376fc7181dc3ca3fbf31378b776bc6ede66b2b50120061544e1cb6bbbae"} Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.791249 4646 generic.go:334] "Generic (PLEG): container finished" podID="85bbeb75-a9c1-4812-916c-53bf850f1cb9" containerID="7f0367103f2eebf9a3973bccc99439fc2683f64f1856bb2eaac28b616dcdab13" exitCode=0 Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.791294 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xg9bs" event={"ID":"85bbeb75-a9c1-4812-916c-53bf850f1cb9","Type":"ContainerDied","Data":"7f0367103f2eebf9a3973bccc99439fc2683f64f1856bb2eaac28b616dcdab13"} Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.791315 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xg9bs" event={"ID":"85bbeb75-a9c1-4812-916c-53bf850f1cb9","Type":"ContainerStarted","Data":"99b79cc335f64683c3b4a84f35c8c279a9ea3bea74901afa34230cb281ef40d4"} Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.793488 4646 generic.go:334] "Generic (PLEG): container finished" podID="237d9726-11ed-4d72-8135-b8ba2f14813f" containerID="560e3e176f09793233b30e3b809ad7455b3c2a1e75a4e698baf8e459815a8207" exitCode=0 Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.793530 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-26vjv" event={"ID":"237d9726-11ed-4d72-8135-b8ba2f14813f","Type":"ContainerDied","Data":"560e3e176f09793233b30e3b809ad7455b3c2a1e75a4e698baf8e459815a8207"} Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.793546 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-26vjv" event={"ID":"237d9726-11ed-4d72-8135-b8ba2f14813f","Type":"ContainerStarted","Data":"45cc4f786bcfb3790fd6883c91b1b53396a654896136cdd9e9e30d63290448b5"} Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.795068 4646 generic.go:334] "Generic (PLEG): container finished" podID="b601410d-7e0e-48e7-b2cb-81126c9ed96b" containerID="9bf0500a838c95634e80cfd84593d031d95f1908712bfc871fb8e341773799f7" exitCode=0 Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.795113 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-11b8-account-create-update-khjqk" event={"ID":"b601410d-7e0e-48e7-b2cb-81126c9ed96b","Type":"ContainerDied","Data":"9bf0500a838c95634e80cfd84593d031d95f1908712bfc871fb8e341773799f7"} Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.795130 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-11b8-account-create-update-khjqk" event={"ID":"b601410d-7e0e-48e7-b2cb-81126c9ed96b","Type":"ContainerStarted","Data":"6769bb2c578a4e73b8e7861191aea01ae90c406e81fd2a74ff843b8ce3135a2c"} Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.796425 4646 generic.go:334] "Generic (PLEG): container finished" podID="22952a9c-263f-40d2-8f5f-54ef029814c8" containerID="4a006684f642fe4d5cc6d6b343aff2dfcf71766df201a33982a877ed08d34ecd" exitCode=0 Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.796459 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9826-account-create-update-9ldlv" event={"ID":"22952a9c-263f-40d2-8f5f-54ef029814c8","Type":"ContainerDied","Data":"4a006684f642fe4d5cc6d6b343aff2dfcf71766df201a33982a877ed08d34ecd"} Dec 03 11:12:01 crc kubenswrapper[4646]: I1203 11:12:01.796474 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9826-account-create-update-9ldlv" event={"ID":"22952a9c-263f-40d2-8f5f-54ef029814c8","Type":"ContainerStarted","Data":"080def12d1a9da22699dc411f781315adae5d7a899878cbe1678e7873a4cf5a9"} Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.236753 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9826-account-create-update-9ldlv" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.329871 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xt6mf\" (UniqueName: \"kubernetes.io/projected/22952a9c-263f-40d2-8f5f-54ef029814c8-kube-api-access-xt6mf\") pod \"22952a9c-263f-40d2-8f5f-54ef029814c8\" (UID: \"22952a9c-263f-40d2-8f5f-54ef029814c8\") " Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.330270 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22952a9c-263f-40d2-8f5f-54ef029814c8-operator-scripts\") pod \"22952a9c-263f-40d2-8f5f-54ef029814c8\" (UID: \"22952a9c-263f-40d2-8f5f-54ef029814c8\") " Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.331288 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22952a9c-263f-40d2-8f5f-54ef029814c8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "22952a9c-263f-40d2-8f5f-54ef029814c8" (UID: "22952a9c-263f-40d2-8f5f-54ef029814c8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.338911 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22952a9c-263f-40d2-8f5f-54ef029814c8-kube-api-access-xt6mf" (OuterVolumeSpecName: "kube-api-access-xt6mf") pod "22952a9c-263f-40d2-8f5f-54ef029814c8" (UID: "22952a9c-263f-40d2-8f5f-54ef029814c8"). InnerVolumeSpecName "kube-api-access-xt6mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.432235 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22952a9c-263f-40d2-8f5f-54ef029814c8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.432270 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xt6mf\" (UniqueName: \"kubernetes.io/projected/22952a9c-263f-40d2-8f5f-54ef029814c8-kube-api-access-xt6mf\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.439139 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-11b8-account-create-update-khjqk" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.444763 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bdf0-account-create-update-z84rm" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.450418 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xg9bs" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.470998 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-26vjv" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.486379 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6grfm" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.532896 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/237d9726-11ed-4d72-8135-b8ba2f14813f-operator-scripts\") pod \"237d9726-11ed-4d72-8135-b8ba2f14813f\" (UID: \"237d9726-11ed-4d72-8135-b8ba2f14813f\") " Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.532969 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b601410d-7e0e-48e7-b2cb-81126c9ed96b-operator-scripts\") pod \"b601410d-7e0e-48e7-b2cb-81126c9ed96b\" (UID: \"b601410d-7e0e-48e7-b2cb-81126c9ed96b\") " Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.532997 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af4fb79d-1d31-4ae4-8769-ff0e6bae317f-operator-scripts\") pod \"af4fb79d-1d31-4ae4-8769-ff0e6bae317f\" (UID: \"af4fb79d-1d31-4ae4-8769-ff0e6bae317f\") " Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.533030 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q74z\" (UniqueName: \"kubernetes.io/projected/af4fb79d-1d31-4ae4-8769-ff0e6bae317f-kube-api-access-7q74z\") pod \"af4fb79d-1d31-4ae4-8769-ff0e6bae317f\" (UID: \"af4fb79d-1d31-4ae4-8769-ff0e6bae317f\") " Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.533063 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85bbeb75-a9c1-4812-916c-53bf850f1cb9-operator-scripts\") pod \"85bbeb75-a9c1-4812-916c-53bf850f1cb9\" (UID: \"85bbeb75-a9c1-4812-916c-53bf850f1cb9\") " Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.533082 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvjg5\" (UniqueName: \"kubernetes.io/projected/237d9726-11ed-4d72-8135-b8ba2f14813f-kube-api-access-gvjg5\") pod \"237d9726-11ed-4d72-8135-b8ba2f14813f\" (UID: \"237d9726-11ed-4d72-8135-b8ba2f14813f\") " Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.533124 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkbht\" (UniqueName: \"kubernetes.io/projected/b601410d-7e0e-48e7-b2cb-81126c9ed96b-kube-api-access-jkbht\") pod \"b601410d-7e0e-48e7-b2cb-81126c9ed96b\" (UID: \"b601410d-7e0e-48e7-b2cb-81126c9ed96b\") " Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.533186 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcknk\" (UniqueName: \"kubernetes.io/projected/85bbeb75-a9c1-4812-916c-53bf850f1cb9-kube-api-access-pcknk\") pod \"85bbeb75-a9c1-4812-916c-53bf850f1cb9\" (UID: \"85bbeb75-a9c1-4812-916c-53bf850f1cb9\") " Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.534409 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85bbeb75-a9c1-4812-916c-53bf850f1cb9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "85bbeb75-a9c1-4812-916c-53bf850f1cb9" (UID: "85bbeb75-a9c1-4812-916c-53bf850f1cb9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.534749 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/237d9726-11ed-4d72-8135-b8ba2f14813f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "237d9726-11ed-4d72-8135-b8ba2f14813f" (UID: "237d9726-11ed-4d72-8135-b8ba2f14813f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.534551 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b601410d-7e0e-48e7-b2cb-81126c9ed96b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b601410d-7e0e-48e7-b2cb-81126c9ed96b" (UID: "b601410d-7e0e-48e7-b2cb-81126c9ed96b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.535394 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af4fb79d-1d31-4ae4-8769-ff0e6bae317f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "af4fb79d-1d31-4ae4-8769-ff0e6bae317f" (UID: "af4fb79d-1d31-4ae4-8769-ff0e6bae317f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.537523 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85bbeb75-a9c1-4812-916c-53bf850f1cb9-kube-api-access-pcknk" (OuterVolumeSpecName: "kube-api-access-pcknk") pod "85bbeb75-a9c1-4812-916c-53bf850f1cb9" (UID: "85bbeb75-a9c1-4812-916c-53bf850f1cb9"). InnerVolumeSpecName "kube-api-access-pcknk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.538254 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/237d9726-11ed-4d72-8135-b8ba2f14813f-kube-api-access-gvjg5" (OuterVolumeSpecName: "kube-api-access-gvjg5") pod "237d9726-11ed-4d72-8135-b8ba2f14813f" (UID: "237d9726-11ed-4d72-8135-b8ba2f14813f"). InnerVolumeSpecName "kube-api-access-gvjg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.539208 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af4fb79d-1d31-4ae4-8769-ff0e6bae317f-kube-api-access-7q74z" (OuterVolumeSpecName: "kube-api-access-7q74z") pod "af4fb79d-1d31-4ae4-8769-ff0e6bae317f" (UID: "af4fb79d-1d31-4ae4-8769-ff0e6bae317f"). InnerVolumeSpecName "kube-api-access-7q74z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.543328 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b601410d-7e0e-48e7-b2cb-81126c9ed96b-kube-api-access-jkbht" (OuterVolumeSpecName: "kube-api-access-jkbht") pod "b601410d-7e0e-48e7-b2cb-81126c9ed96b" (UID: "b601410d-7e0e-48e7-b2cb-81126c9ed96b"). InnerVolumeSpecName "kube-api-access-jkbht". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.635212 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjn54\" (UniqueName: \"kubernetes.io/projected/b0660b75-6739-4a9e-8cb4-ff6b15b0080f-kube-api-access-sjn54\") pod \"b0660b75-6739-4a9e-8cb4-ff6b15b0080f\" (UID: \"b0660b75-6739-4a9e-8cb4-ff6b15b0080f\") " Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.635987 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0660b75-6739-4a9e-8cb4-ff6b15b0080f-operator-scripts\") pod \"b0660b75-6739-4a9e-8cb4-ff6b15b0080f\" (UID: \"b0660b75-6739-4a9e-8cb4-ff6b15b0080f\") " Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.636421 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkbht\" (UniqueName: \"kubernetes.io/projected/b601410d-7e0e-48e7-b2cb-81126c9ed96b-kube-api-access-jkbht\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.636440 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcknk\" (UniqueName: \"kubernetes.io/projected/85bbeb75-a9c1-4812-916c-53bf850f1cb9-kube-api-access-pcknk\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.636452 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/237d9726-11ed-4d72-8135-b8ba2f14813f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.636462 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b601410d-7e0e-48e7-b2cb-81126c9ed96b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.636472 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af4fb79d-1d31-4ae4-8769-ff0e6bae317f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.636482 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q74z\" (UniqueName: \"kubernetes.io/projected/af4fb79d-1d31-4ae4-8769-ff0e6bae317f-kube-api-access-7q74z\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.636492 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85bbeb75-a9c1-4812-916c-53bf850f1cb9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.636504 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvjg5\" (UniqueName: \"kubernetes.io/projected/237d9726-11ed-4d72-8135-b8ba2f14813f-kube-api-access-gvjg5\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.636886 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0660b75-6739-4a9e-8cb4-ff6b15b0080f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b0660b75-6739-4a9e-8cb4-ff6b15b0080f" (UID: "b0660b75-6739-4a9e-8cb4-ff6b15b0080f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.639781 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0660b75-6739-4a9e-8cb4-ff6b15b0080f-kube-api-access-sjn54" (OuterVolumeSpecName: "kube-api-access-sjn54") pod "b0660b75-6739-4a9e-8cb4-ff6b15b0080f" (UID: "b0660b75-6739-4a9e-8cb4-ff6b15b0080f"). InnerVolumeSpecName "kube-api-access-sjn54". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.738509 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0660b75-6739-4a9e-8cb4-ff6b15b0080f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.738548 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjn54\" (UniqueName: \"kubernetes.io/projected/b0660b75-6739-4a9e-8cb4-ff6b15b0080f-kube-api-access-sjn54\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.822878 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xg9bs" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.822912 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xg9bs" event={"ID":"85bbeb75-a9c1-4812-916c-53bf850f1cb9","Type":"ContainerDied","Data":"99b79cc335f64683c3b4a84f35c8c279a9ea3bea74901afa34230cb281ef40d4"} Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.822970 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99b79cc335f64683c3b4a84f35c8c279a9ea3bea74901afa34230cb281ef40d4" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.824539 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-26vjv" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.824534 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-26vjv" event={"ID":"237d9726-11ed-4d72-8135-b8ba2f14813f","Type":"ContainerDied","Data":"45cc4f786bcfb3790fd6883c91b1b53396a654896136cdd9e9e30d63290448b5"} Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.824675 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45cc4f786bcfb3790fd6883c91b1b53396a654896136cdd9e9e30d63290448b5" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.825692 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-11b8-account-create-update-khjqk" event={"ID":"b601410d-7e0e-48e7-b2cb-81126c9ed96b","Type":"ContainerDied","Data":"6769bb2c578a4e73b8e7861191aea01ae90c406e81fd2a74ff843b8ce3135a2c"} Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.825721 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6769bb2c578a4e73b8e7861191aea01ae90c406e81fd2a74ff843b8ce3135a2c" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.825763 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-11b8-account-create-update-khjqk" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.829873 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9826-account-create-update-9ldlv" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.829880 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9826-account-create-update-9ldlv" event={"ID":"22952a9c-263f-40d2-8f5f-54ef029814c8","Type":"ContainerDied","Data":"080def12d1a9da22699dc411f781315adae5d7a899878cbe1678e7873a4cf5a9"} Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.829909 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="080def12d1a9da22699dc411f781315adae5d7a899878cbe1678e7873a4cf5a9" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.830948 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bdf0-account-create-update-z84rm" event={"ID":"af4fb79d-1d31-4ae4-8769-ff0e6bae317f","Type":"ContainerDied","Data":"d2da5ab1eca3e1964302d27e53499e0181485c15498e379c738aaa10fafe0f1c"} Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.830963 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bdf0-account-create-update-z84rm" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.830968 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2da5ab1eca3e1964302d27e53499e0181485c15498e379c738aaa10fafe0f1c" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.831850 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6grfm" event={"ID":"b0660b75-6739-4a9e-8cb4-ff6b15b0080f","Type":"ContainerDied","Data":"935b5376fc7181dc3ca3fbf31378b776bc6ede66b2b50120061544e1cb6bbbae"} Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.831868 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="935b5376fc7181dc3ca3fbf31378b776bc6ede66b2b50120061544e1cb6bbbae" Dec 03 11:12:03 crc kubenswrapper[4646]: I1203 11:12:03.831911 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6grfm" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.661437 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-tjvwv"] Dec 03 11:12:05 crc kubenswrapper[4646]: E1203 11:12:05.662107 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85bbeb75-a9c1-4812-916c-53bf850f1cb9" containerName="mariadb-database-create" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.662122 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="85bbeb75-a9c1-4812-916c-53bf850f1cb9" containerName="mariadb-database-create" Dec 03 11:12:05 crc kubenswrapper[4646]: E1203 11:12:05.662138 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af4fb79d-1d31-4ae4-8769-ff0e6bae317f" containerName="mariadb-account-create-update" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.662145 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="af4fb79d-1d31-4ae4-8769-ff0e6bae317f" containerName="mariadb-account-create-update" Dec 03 11:12:05 crc kubenswrapper[4646]: E1203 11:12:05.662172 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22952a9c-263f-40d2-8f5f-54ef029814c8" containerName="mariadb-account-create-update" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.662180 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="22952a9c-263f-40d2-8f5f-54ef029814c8" containerName="mariadb-account-create-update" Dec 03 11:12:05 crc kubenswrapper[4646]: E1203 11:12:05.662192 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="237d9726-11ed-4d72-8135-b8ba2f14813f" containerName="mariadb-database-create" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.662199 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="237d9726-11ed-4d72-8135-b8ba2f14813f" containerName="mariadb-database-create" Dec 03 11:12:05 crc kubenswrapper[4646]: E1203 11:12:05.662214 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b601410d-7e0e-48e7-b2cb-81126c9ed96b" containerName="mariadb-account-create-update" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.662223 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="b601410d-7e0e-48e7-b2cb-81126c9ed96b" containerName="mariadb-account-create-update" Dec 03 11:12:05 crc kubenswrapper[4646]: E1203 11:12:05.662237 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0660b75-6739-4a9e-8cb4-ff6b15b0080f" containerName="mariadb-database-create" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.662244 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0660b75-6739-4a9e-8cb4-ff6b15b0080f" containerName="mariadb-database-create" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.662432 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="237d9726-11ed-4d72-8135-b8ba2f14813f" containerName="mariadb-database-create" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.662447 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="85bbeb75-a9c1-4812-916c-53bf850f1cb9" containerName="mariadb-database-create" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.662459 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="af4fb79d-1d31-4ae4-8769-ff0e6bae317f" containerName="mariadb-account-create-update" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.662472 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="b601410d-7e0e-48e7-b2cb-81126c9ed96b" containerName="mariadb-account-create-update" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.662490 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="22952a9c-263f-40d2-8f5f-54ef029814c8" containerName="mariadb-account-create-update" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.662502 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0660b75-6739-4a9e-8cb4-ff6b15b0080f" containerName="mariadb-database-create" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.663140 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.665355 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5xl2d" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.665685 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.679595 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tjvwv"] Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.769800 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpfz8\" (UniqueName: \"kubernetes.io/projected/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-kube-api-access-mpfz8\") pod \"glance-db-sync-tjvwv\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.769867 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-config-data\") pod \"glance-db-sync-tjvwv\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.769890 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-db-sync-config-data\") pod \"glance-db-sync-tjvwv\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.769961 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-combined-ca-bundle\") pod \"glance-db-sync-tjvwv\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.782601 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-rjmpb" podUID="f7606fd8-32a7-402e-bac9-bfe718af9019" containerName="ovn-controller" probeResult="failure" output=< Dec 03 11:12:05 crc kubenswrapper[4646]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 11:12:05 crc kubenswrapper[4646]: > Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.790974 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.800133 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-5v9pv" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.859388 4646 generic.go:334] "Generic (PLEG): container finished" podID="26f0a162-8af0-438b-b48e-5bb76cb9a645" containerID="403c513e6091509f7524e462220d0d7d652a7b5bfdbdd96c35cd55a4ea200891" exitCode=0 Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.860189 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"26f0a162-8af0-438b-b48e-5bb76cb9a645","Type":"ContainerDied","Data":"403c513e6091509f7524e462220d0d7d652a7b5bfdbdd96c35cd55a4ea200891"} Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.872027 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-config-data\") pod \"glance-db-sync-tjvwv\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.873059 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-db-sync-config-data\") pod \"glance-db-sync-tjvwv\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.873146 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-combined-ca-bundle\") pod \"glance-db-sync-tjvwv\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.873280 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpfz8\" (UniqueName: \"kubernetes.io/projected/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-kube-api-access-mpfz8\") pod \"glance-db-sync-tjvwv\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.876695 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-config-data\") pod \"glance-db-sync-tjvwv\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.879723 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-db-sync-config-data\") pod \"glance-db-sync-tjvwv\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.881508 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-combined-ca-bundle\") pod \"glance-db-sync-tjvwv\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:05 crc kubenswrapper[4646]: I1203 11:12:05.893782 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpfz8\" (UniqueName: \"kubernetes.io/projected/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-kube-api-access-mpfz8\") pod \"glance-db-sync-tjvwv\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.014527 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.047842 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rjmpb-config-vftxg"] Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.050472 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.053959 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.115303 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rjmpb-config-vftxg"] Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.180399 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-log-ovn\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.180604 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-run\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.180639 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-run-ovn\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.180669 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/561b06c0-28a3-449c-838f-31e8cb42efb7-additional-scripts\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.180695 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwbh5\" (UniqueName: \"kubernetes.io/projected/561b06c0-28a3-449c-838f-31e8cb42efb7-kube-api-access-wwbh5\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.180713 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/561b06c0-28a3-449c-838f-31e8cb42efb7-scripts\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.281921 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/561b06c0-28a3-449c-838f-31e8cb42efb7-additional-scripts\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.282921 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/561b06c0-28a3-449c-838f-31e8cb42efb7-additional-scripts\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.283000 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwbh5\" (UniqueName: \"kubernetes.io/projected/561b06c0-28a3-449c-838f-31e8cb42efb7-kube-api-access-wwbh5\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.283037 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/561b06c0-28a3-449c-838f-31e8cb42efb7-scripts\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.283171 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-log-ovn\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.283205 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-run\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.283252 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-run-ovn\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.283987 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-run\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.283958 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-log-ovn\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.284000 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-run-ovn\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.285504 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/561b06c0-28a3-449c-838f-31e8cb42efb7-scripts\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.304148 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwbh5\" (UniqueName: \"kubernetes.io/projected/561b06c0-28a3-449c-838f-31e8cb42efb7-kube-api-access-wwbh5\") pod \"ovn-controller-rjmpb-config-vftxg\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.417454 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.690352 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tjvwv"] Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.867752 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tjvwv" event={"ID":"f23e454b-e13d-47f0-a3ba-2e96f3cc0649","Type":"ContainerStarted","Data":"0322f616c0e6df2162e21879f2b3bd99c1066db369fabfda898d47f01f9b2c6d"} Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.869952 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"26f0a162-8af0-438b-b48e-5bb76cb9a645","Type":"ContainerStarted","Data":"98c8221e6383dc19e35b0d441fe752daa788c1ed0ed5b4a1b5eb71276729f617"} Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.870753 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.915596 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371954.939201 podStartE2EDuration="1m21.915575091s" podCreationTimestamp="2025-12-03 11:10:45 +0000 UTC" firstStartedPulling="2025-12-03 11:10:47.510510266 +0000 UTC m=+1023.973566401" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:12:06.905286456 +0000 UTC m=+1103.368342601" watchObservedRunningTime="2025-12-03 11:12:06.915575091 +0000 UTC m=+1103.378631226" Dec 03 11:12:06 crc kubenswrapper[4646]: I1203 11:12:06.943778 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rjmpb-config-vftxg"] Dec 03 11:12:07 crc kubenswrapper[4646]: I1203 11:12:07.306602 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:12:07 crc kubenswrapper[4646]: I1203 11:12:07.901905 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rjmpb-config-vftxg" event={"ID":"561b06c0-28a3-449c-838f-31e8cb42efb7","Type":"ContainerStarted","Data":"6a79067ebbc0d9ee89292863e21922a5321ab9795bdca740f490c970db84a714"} Dec 03 11:12:07 crc kubenswrapper[4646]: I1203 11:12:07.901949 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rjmpb-config-vftxg" event={"ID":"561b06c0-28a3-449c-838f-31e8cb42efb7","Type":"ContainerStarted","Data":"cb63a875f6ce2e95bd08af7bd2ca41ab61ecf3160cc7ff0d78c98a8e9b238dcb"} Dec 03 11:12:08 crc kubenswrapper[4646]: I1203 11:12:08.931160 4646 generic.go:334] "Generic (PLEG): container finished" podID="561b06c0-28a3-449c-838f-31e8cb42efb7" containerID="6a79067ebbc0d9ee89292863e21922a5321ab9795bdca740f490c970db84a714" exitCode=0 Dec 03 11:12:08 crc kubenswrapper[4646]: I1203 11:12:08.931554 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rjmpb-config-vftxg" event={"ID":"561b06c0-28a3-449c-838f-31e8cb42efb7","Type":"ContainerDied","Data":"6a79067ebbc0d9ee89292863e21922a5321ab9795bdca740f490c970db84a714"} Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.290962 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.380499 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwbh5\" (UniqueName: \"kubernetes.io/projected/561b06c0-28a3-449c-838f-31e8cb42efb7-kube-api-access-wwbh5\") pod \"561b06c0-28a3-449c-838f-31e8cb42efb7\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.380857 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/561b06c0-28a3-449c-838f-31e8cb42efb7-scripts\") pod \"561b06c0-28a3-449c-838f-31e8cb42efb7\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.380976 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-run\") pod \"561b06c0-28a3-449c-838f-31e8cb42efb7\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.381035 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-log-ovn\") pod \"561b06c0-28a3-449c-838f-31e8cb42efb7\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.381055 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/561b06c0-28a3-449c-838f-31e8cb42efb7-additional-scripts\") pod \"561b06c0-28a3-449c-838f-31e8cb42efb7\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.381053 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-run" (OuterVolumeSpecName: "var-run") pod "561b06c0-28a3-449c-838f-31e8cb42efb7" (UID: "561b06c0-28a3-449c-838f-31e8cb42efb7"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.381081 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-run-ovn\") pod \"561b06c0-28a3-449c-838f-31e8cb42efb7\" (UID: \"561b06c0-28a3-449c-838f-31e8cb42efb7\") " Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.381100 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "561b06c0-28a3-449c-838f-31e8cb42efb7" (UID: "561b06c0-28a3-449c-838f-31e8cb42efb7"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.381236 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "561b06c0-28a3-449c-838f-31e8cb42efb7" (UID: "561b06c0-28a3-449c-838f-31e8cb42efb7"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.381551 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/561b06c0-28a3-449c-838f-31e8cb42efb7-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "561b06c0-28a3-449c-838f-31e8cb42efb7" (UID: "561b06c0-28a3-449c-838f-31e8cb42efb7"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.381770 4646 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.381794 4646 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.381806 4646 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/561b06c0-28a3-449c-838f-31e8cb42efb7-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.381816 4646 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/561b06c0-28a3-449c-838f-31e8cb42efb7-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.382300 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/561b06c0-28a3-449c-838f-31e8cb42efb7-scripts" (OuterVolumeSpecName: "scripts") pod "561b06c0-28a3-449c-838f-31e8cb42efb7" (UID: "561b06c0-28a3-449c-838f-31e8cb42efb7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.394980 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/561b06c0-28a3-449c-838f-31e8cb42efb7-kube-api-access-wwbh5" (OuterVolumeSpecName: "kube-api-access-wwbh5") pod "561b06c0-28a3-449c-838f-31e8cb42efb7" (UID: "561b06c0-28a3-449c-838f-31e8cb42efb7"). InnerVolumeSpecName "kube-api-access-wwbh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.483177 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwbh5\" (UniqueName: \"kubernetes.io/projected/561b06c0-28a3-449c-838f-31e8cb42efb7-kube-api-access-wwbh5\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.483234 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/561b06c0-28a3-449c-838f-31e8cb42efb7-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.791928 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-rjmpb" Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.947397 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rjmpb-config-vftxg" event={"ID":"561b06c0-28a3-449c-838f-31e8cb42efb7","Type":"ContainerDied","Data":"cb63a875f6ce2e95bd08af7bd2ca41ab61ecf3160cc7ff0d78c98a8e9b238dcb"} Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.947434 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb63a875f6ce2e95bd08af7bd2ca41ab61ecf3160cc7ff0d78c98a8e9b238dcb" Dec 03 11:12:10 crc kubenswrapper[4646]: I1203 11:12:10.947481 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rjmpb-config-vftxg" Dec 03 11:12:11 crc kubenswrapper[4646]: I1203 11:12:11.100736 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rjmpb-config-vftxg"] Dec 03 11:12:11 crc kubenswrapper[4646]: I1203 11:12:11.110110 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-rjmpb-config-vftxg"] Dec 03 11:12:11 crc kubenswrapper[4646]: I1203 11:12:11.859920 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="561b06c0-28a3-449c-838f-31e8cb42efb7" path="/var/lib/kubelet/pods/561b06c0-28a3-449c-838f-31e8cb42efb7/volumes" Dec 03 11:12:16 crc kubenswrapper[4646]: I1203 11:12:16.807564 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.251046 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-d2vp8"] Dec 03 11:12:17 crc kubenswrapper[4646]: E1203 11:12:17.251455 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561b06c0-28a3-449c-838f-31e8cb42efb7" containerName="ovn-config" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.251473 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="561b06c0-28a3-449c-838f-31e8cb42efb7" containerName="ovn-config" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.251633 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="561b06c0-28a3-449c-838f-31e8cb42efb7" containerName="ovn-config" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.252271 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d2vp8" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.260576 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-5129-account-create-update-drkql"] Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.261593 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5129-account-create-update-drkql" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.263768 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.272193 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-d2vp8"] Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.280251 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5129-account-create-update-drkql"] Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.369562 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-e410-account-create-update-2nflx"] Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.370909 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e410-account-create-update-2nflx" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.375901 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.390489 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-d95tn"] Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.392056 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-d95tn" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.400441 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-e410-account-create-update-2nflx"] Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.403220 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b26f36df-eade-45a7-84e5-a181b4b3d43f-operator-scripts\") pod \"barbican-db-create-d2vp8\" (UID: \"b26f36df-eade-45a7-84e5-a181b4b3d43f\") " pod="openstack/barbican-db-create-d2vp8" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.403326 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5b8e5a9-ea00-4d58-8e49-50c72b9679de-operator-scripts\") pod \"barbican-5129-account-create-update-drkql\" (UID: \"b5b8e5a9-ea00-4d58-8e49-50c72b9679de\") " pod="openstack/barbican-5129-account-create-update-drkql" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.403396 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jts9\" (UniqueName: \"kubernetes.io/projected/b5b8e5a9-ea00-4d58-8e49-50c72b9679de-kube-api-access-8jts9\") pod \"barbican-5129-account-create-update-drkql\" (UID: \"b5b8e5a9-ea00-4d58-8e49-50c72b9679de\") " pod="openstack/barbican-5129-account-create-update-drkql" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.403423 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmxvt\" (UniqueName: \"kubernetes.io/projected/b26f36df-eade-45a7-84e5-a181b4b3d43f-kube-api-access-gmxvt\") pod \"barbican-db-create-d2vp8\" (UID: \"b26f36df-eade-45a7-84e5-a181b4b3d43f\") " pod="openstack/barbican-db-create-d2vp8" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.411995 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-d95tn"] Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.505149 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jts9\" (UniqueName: \"kubernetes.io/projected/b5b8e5a9-ea00-4d58-8e49-50c72b9679de-kube-api-access-8jts9\") pod \"barbican-5129-account-create-update-drkql\" (UID: \"b5b8e5a9-ea00-4d58-8e49-50c72b9679de\") " pod="openstack/barbican-5129-account-create-update-drkql" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.505194 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmxvt\" (UniqueName: \"kubernetes.io/projected/b26f36df-eade-45a7-84e5-a181b4b3d43f-kube-api-access-gmxvt\") pod \"barbican-db-create-d2vp8\" (UID: \"b26f36df-eade-45a7-84e5-a181b4b3d43f\") " pod="openstack/barbican-db-create-d2vp8" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.505219 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh6pj\" (UniqueName: \"kubernetes.io/projected/dc578e46-8ab2-4567-91ba-66462a04a328-kube-api-access-lh6pj\") pod \"cinder-e410-account-create-update-2nflx\" (UID: \"dc578e46-8ab2-4567-91ba-66462a04a328\") " pod="openstack/cinder-e410-account-create-update-2nflx" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.505259 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76e53db6-897c-4c62-b732-034a119ab559-operator-scripts\") pod \"cinder-db-create-d95tn\" (UID: \"76e53db6-897c-4c62-b732-034a119ab559\") " pod="openstack/cinder-db-create-d95tn" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.505289 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5kdf\" (UniqueName: \"kubernetes.io/projected/76e53db6-897c-4c62-b732-034a119ab559-kube-api-access-z5kdf\") pod \"cinder-db-create-d95tn\" (UID: \"76e53db6-897c-4c62-b732-034a119ab559\") " pod="openstack/cinder-db-create-d95tn" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.505322 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b26f36df-eade-45a7-84e5-a181b4b3d43f-operator-scripts\") pod \"barbican-db-create-d2vp8\" (UID: \"b26f36df-eade-45a7-84e5-a181b4b3d43f\") " pod="openstack/barbican-db-create-d2vp8" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.505390 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc578e46-8ab2-4567-91ba-66462a04a328-operator-scripts\") pod \"cinder-e410-account-create-update-2nflx\" (UID: \"dc578e46-8ab2-4567-91ba-66462a04a328\") " pod="openstack/cinder-e410-account-create-update-2nflx" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.505407 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5b8e5a9-ea00-4d58-8e49-50c72b9679de-operator-scripts\") pod \"barbican-5129-account-create-update-drkql\" (UID: \"b5b8e5a9-ea00-4d58-8e49-50c72b9679de\") " pod="openstack/barbican-5129-account-create-update-drkql" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.506193 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5b8e5a9-ea00-4d58-8e49-50c72b9679de-operator-scripts\") pod \"barbican-5129-account-create-update-drkql\" (UID: \"b5b8e5a9-ea00-4d58-8e49-50c72b9679de\") " pod="openstack/barbican-5129-account-create-update-drkql" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.506584 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b26f36df-eade-45a7-84e5-a181b4b3d43f-operator-scripts\") pod \"barbican-db-create-d2vp8\" (UID: \"b26f36df-eade-45a7-84e5-a181b4b3d43f\") " pod="openstack/barbican-db-create-d2vp8" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.527286 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmxvt\" (UniqueName: \"kubernetes.io/projected/b26f36df-eade-45a7-84e5-a181b4b3d43f-kube-api-access-gmxvt\") pod \"barbican-db-create-d2vp8\" (UID: \"b26f36df-eade-45a7-84e5-a181b4b3d43f\") " pod="openstack/barbican-db-create-d2vp8" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.527359 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jts9\" (UniqueName: \"kubernetes.io/projected/b5b8e5a9-ea00-4d58-8e49-50c72b9679de-kube-api-access-8jts9\") pod \"barbican-5129-account-create-update-drkql\" (UID: \"b5b8e5a9-ea00-4d58-8e49-50c72b9679de\") " pod="openstack/barbican-5129-account-create-update-drkql" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.561733 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-jpkbq"] Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.563243 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jpkbq" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.567371 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.569826 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.570029 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ddjxq" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.570378 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.573449 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d2vp8" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.584434 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5129-account-create-update-drkql" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.607194 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76e53db6-897c-4c62-b732-034a119ab559-operator-scripts\") pod \"cinder-db-create-d95tn\" (UID: \"76e53db6-897c-4c62-b732-034a119ab559\") " pod="openstack/cinder-db-create-d95tn" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.607267 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5kdf\" (UniqueName: \"kubernetes.io/projected/76e53db6-897c-4c62-b732-034a119ab559-kube-api-access-z5kdf\") pod \"cinder-db-create-d95tn\" (UID: \"76e53db6-897c-4c62-b732-034a119ab559\") " pod="openstack/cinder-db-create-d95tn" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.607369 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc578e46-8ab2-4567-91ba-66462a04a328-operator-scripts\") pod \"cinder-e410-account-create-update-2nflx\" (UID: \"dc578e46-8ab2-4567-91ba-66462a04a328\") " pod="openstack/cinder-e410-account-create-update-2nflx" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.607412 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh6pj\" (UniqueName: \"kubernetes.io/projected/dc578e46-8ab2-4567-91ba-66462a04a328-kube-api-access-lh6pj\") pod \"cinder-e410-account-create-update-2nflx\" (UID: \"dc578e46-8ab2-4567-91ba-66462a04a328\") " pod="openstack/cinder-e410-account-create-update-2nflx" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.608761 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76e53db6-897c-4c62-b732-034a119ab559-operator-scripts\") pod \"cinder-db-create-d95tn\" (UID: \"76e53db6-897c-4c62-b732-034a119ab559\") " pod="openstack/cinder-db-create-d95tn" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.609244 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc578e46-8ab2-4567-91ba-66462a04a328-operator-scripts\") pod \"cinder-e410-account-create-update-2nflx\" (UID: \"dc578e46-8ab2-4567-91ba-66462a04a328\") " pod="openstack/cinder-e410-account-create-update-2nflx" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.632131 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jpkbq"] Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.674047 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-w557d"] Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.675252 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-w557d" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.697830 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-w557d"] Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.700381 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5kdf\" (UniqueName: \"kubernetes.io/projected/76e53db6-897c-4c62-b732-034a119ab559-kube-api-access-z5kdf\") pod \"cinder-db-create-d95tn\" (UID: \"76e53db6-897c-4c62-b732-034a119ab559\") " pod="openstack/cinder-db-create-d95tn" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.714015 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1785e265-f0f0-4e8f-8185-8f84b3121f3b-combined-ca-bundle\") pod \"keystone-db-sync-jpkbq\" (UID: \"1785e265-f0f0-4e8f-8185-8f84b3121f3b\") " pod="openstack/keystone-db-sync-jpkbq" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.714102 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1785e265-f0f0-4e8f-8185-8f84b3121f3b-config-data\") pod \"keystone-db-sync-jpkbq\" (UID: \"1785e265-f0f0-4e8f-8185-8f84b3121f3b\") " pod="openstack/keystone-db-sync-jpkbq" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.714130 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffx6l\" (UniqueName: \"kubernetes.io/projected/1785e265-f0f0-4e8f-8185-8f84b3121f3b-kube-api-access-ffx6l\") pod \"keystone-db-sync-jpkbq\" (UID: \"1785e265-f0f0-4e8f-8185-8f84b3121f3b\") " pod="openstack/keystone-db-sync-jpkbq" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.714940 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh6pj\" (UniqueName: \"kubernetes.io/projected/dc578e46-8ab2-4567-91ba-66462a04a328-kube-api-access-lh6pj\") pod \"cinder-e410-account-create-update-2nflx\" (UID: \"dc578e46-8ab2-4567-91ba-66462a04a328\") " pod="openstack/cinder-e410-account-create-update-2nflx" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.722392 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-9dde-account-create-update-7ch8v"] Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.723543 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9dde-account-create-update-7ch8v" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.730511 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-d95tn" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.735207 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.815674 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1785e265-f0f0-4e8f-8185-8f84b3121f3b-config-data\") pod \"keystone-db-sync-jpkbq\" (UID: \"1785e265-f0f0-4e8f-8185-8f84b3121f3b\") " pod="openstack/keystone-db-sync-jpkbq" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.815754 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffx6l\" (UniqueName: \"kubernetes.io/projected/1785e265-f0f0-4e8f-8185-8f84b3121f3b-kube-api-access-ffx6l\") pod \"keystone-db-sync-jpkbq\" (UID: \"1785e265-f0f0-4e8f-8185-8f84b3121f3b\") " pod="openstack/keystone-db-sync-jpkbq" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.815804 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cec4fe45-3c8d-4dea-b5af-7fe6cff4594d-operator-scripts\") pod \"neutron-db-create-w557d\" (UID: \"cec4fe45-3c8d-4dea-b5af-7fe6cff4594d\") " pod="openstack/neutron-db-create-w557d" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.815882 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1785e265-f0f0-4e8f-8185-8f84b3121f3b-combined-ca-bundle\") pod \"keystone-db-sync-jpkbq\" (UID: \"1785e265-f0f0-4e8f-8185-8f84b3121f3b\") " pod="openstack/keystone-db-sync-jpkbq" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.815908 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrw4t\" (UniqueName: \"kubernetes.io/projected/cec4fe45-3c8d-4dea-b5af-7fe6cff4594d-kube-api-access-rrw4t\") pod \"neutron-db-create-w557d\" (UID: \"cec4fe45-3c8d-4dea-b5af-7fe6cff4594d\") " pod="openstack/neutron-db-create-w557d" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.819276 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1785e265-f0f0-4e8f-8185-8f84b3121f3b-config-data\") pod \"keystone-db-sync-jpkbq\" (UID: \"1785e265-f0f0-4e8f-8185-8f84b3121f3b\") " pod="openstack/keystone-db-sync-jpkbq" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.834946 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1785e265-f0f0-4e8f-8185-8f84b3121f3b-combined-ca-bundle\") pod \"keystone-db-sync-jpkbq\" (UID: \"1785e265-f0f0-4e8f-8185-8f84b3121f3b\") " pod="openstack/keystone-db-sync-jpkbq" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.842277 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffx6l\" (UniqueName: \"kubernetes.io/projected/1785e265-f0f0-4e8f-8185-8f84b3121f3b-kube-api-access-ffx6l\") pod \"keystone-db-sync-jpkbq\" (UID: \"1785e265-f0f0-4e8f-8185-8f84b3121f3b\") " pod="openstack/keystone-db-sync-jpkbq" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.865171 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9dde-account-create-update-7ch8v"] Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.916902 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jpkbq" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.917507 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lt2r\" (UniqueName: \"kubernetes.io/projected/16b24e6b-5a24-4abd-b3de-7932ce9e63c3-kube-api-access-6lt2r\") pod \"neutron-9dde-account-create-update-7ch8v\" (UID: \"16b24e6b-5a24-4abd-b3de-7932ce9e63c3\") " pod="openstack/neutron-9dde-account-create-update-7ch8v" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.917585 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrw4t\" (UniqueName: \"kubernetes.io/projected/cec4fe45-3c8d-4dea-b5af-7fe6cff4594d-kube-api-access-rrw4t\") pod \"neutron-db-create-w557d\" (UID: \"cec4fe45-3c8d-4dea-b5af-7fe6cff4594d\") " pod="openstack/neutron-db-create-w557d" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.917653 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16b24e6b-5a24-4abd-b3de-7932ce9e63c3-operator-scripts\") pod \"neutron-9dde-account-create-update-7ch8v\" (UID: \"16b24e6b-5a24-4abd-b3de-7932ce9e63c3\") " pod="openstack/neutron-9dde-account-create-update-7ch8v" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.917716 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cec4fe45-3c8d-4dea-b5af-7fe6cff4594d-operator-scripts\") pod \"neutron-db-create-w557d\" (UID: \"cec4fe45-3c8d-4dea-b5af-7fe6cff4594d\") " pod="openstack/neutron-db-create-w557d" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.918555 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cec4fe45-3c8d-4dea-b5af-7fe6cff4594d-operator-scripts\") pod \"neutron-db-create-w557d\" (UID: \"cec4fe45-3c8d-4dea-b5af-7fe6cff4594d\") " pod="openstack/neutron-db-create-w557d" Dec 03 11:12:17 crc kubenswrapper[4646]: I1203 11:12:17.944917 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrw4t\" (UniqueName: \"kubernetes.io/projected/cec4fe45-3c8d-4dea-b5af-7fe6cff4594d-kube-api-access-rrw4t\") pod \"neutron-db-create-w557d\" (UID: \"cec4fe45-3c8d-4dea-b5af-7fe6cff4594d\") " pod="openstack/neutron-db-create-w557d" Dec 03 11:12:18 crc kubenswrapper[4646]: I1203 11:12:18.002241 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e410-account-create-update-2nflx" Dec 03 11:12:18 crc kubenswrapper[4646]: I1203 11:12:18.004646 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-w557d" Dec 03 11:12:18 crc kubenswrapper[4646]: I1203 11:12:18.020236 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16b24e6b-5a24-4abd-b3de-7932ce9e63c3-operator-scripts\") pod \"neutron-9dde-account-create-update-7ch8v\" (UID: \"16b24e6b-5a24-4abd-b3de-7932ce9e63c3\") " pod="openstack/neutron-9dde-account-create-update-7ch8v" Dec 03 11:12:18 crc kubenswrapper[4646]: I1203 11:12:18.020434 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lt2r\" (UniqueName: \"kubernetes.io/projected/16b24e6b-5a24-4abd-b3de-7932ce9e63c3-kube-api-access-6lt2r\") pod \"neutron-9dde-account-create-update-7ch8v\" (UID: \"16b24e6b-5a24-4abd-b3de-7932ce9e63c3\") " pod="openstack/neutron-9dde-account-create-update-7ch8v" Dec 03 11:12:18 crc kubenswrapper[4646]: I1203 11:12:18.021813 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16b24e6b-5a24-4abd-b3de-7932ce9e63c3-operator-scripts\") pod \"neutron-9dde-account-create-update-7ch8v\" (UID: \"16b24e6b-5a24-4abd-b3de-7932ce9e63c3\") " pod="openstack/neutron-9dde-account-create-update-7ch8v" Dec 03 11:12:18 crc kubenswrapper[4646]: I1203 11:12:18.061228 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lt2r\" (UniqueName: \"kubernetes.io/projected/16b24e6b-5a24-4abd-b3de-7932ce9e63c3-kube-api-access-6lt2r\") pod \"neutron-9dde-account-create-update-7ch8v\" (UID: \"16b24e6b-5a24-4abd-b3de-7932ce9e63c3\") " pod="openstack/neutron-9dde-account-create-update-7ch8v" Dec 03 11:12:18 crc kubenswrapper[4646]: I1203 11:12:18.102282 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9dde-account-create-update-7ch8v" Dec 03 11:12:22 crc kubenswrapper[4646]: I1203 11:12:22.319636 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jpkbq"] Dec 03 11:12:22 crc kubenswrapper[4646]: I1203 11:12:22.402979 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-d2vp8"] Dec 03 11:12:22 crc kubenswrapper[4646]: W1203 11:12:22.418770 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb26f36df_eade_45a7_84e5_a181b4b3d43f.slice/crio-ecc17f38e541d7e7ae5fe93430da9482c78ccc94583d79e4d6be56db3467ab0f WatchSource:0}: Error finding container ecc17f38e541d7e7ae5fe93430da9482c78ccc94583d79e4d6be56db3467ab0f: Status 404 returned error can't find the container with id ecc17f38e541d7e7ae5fe93430da9482c78ccc94583d79e4d6be56db3467ab0f Dec 03 11:12:22 crc kubenswrapper[4646]: W1203 11:12:22.460787 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc578e46_8ab2_4567_91ba_66462a04a328.slice/crio-67f621298ca4faa316d821e9578986def8e82232eae9d8dc1f32033ce299f158 WatchSource:0}: Error finding container 67f621298ca4faa316d821e9578986def8e82232eae9d8dc1f32033ce299f158: Status 404 returned error can't find the container with id 67f621298ca4faa316d821e9578986def8e82232eae9d8dc1f32033ce299f158 Dec 03 11:12:22 crc kubenswrapper[4646]: I1203 11:12:22.461004 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9dde-account-create-update-7ch8v"] Dec 03 11:12:22 crc kubenswrapper[4646]: W1203 11:12:22.462611 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16b24e6b_5a24_4abd_b3de_7932ce9e63c3.slice/crio-f13a37fcec8f4e25176c042155a9a3d397c1d203d7007c7e0809768edfd3ef6e WatchSource:0}: Error finding container f13a37fcec8f4e25176c042155a9a3d397c1d203d7007c7e0809768edfd3ef6e: Status 404 returned error can't find the container with id f13a37fcec8f4e25176c042155a9a3d397c1d203d7007c7e0809768edfd3ef6e Dec 03 11:12:22 crc kubenswrapper[4646]: I1203 11:12:22.469120 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-e410-account-create-update-2nflx"] Dec 03 11:12:22 crc kubenswrapper[4646]: I1203 11:12:22.479526 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-d95tn"] Dec 03 11:12:22 crc kubenswrapper[4646]: I1203 11:12:22.597697 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5129-account-create-update-drkql"] Dec 03 11:12:22 crc kubenswrapper[4646]: I1203 11:12:22.607715 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-w557d"] Dec 03 11:12:22 crc kubenswrapper[4646]: W1203 11:12:22.618984 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5b8e5a9_ea00_4d58_8e49_50c72b9679de.slice/crio-ab924ef9bf311922a6ae1b8afad828fab05fec88ddd13d6a38d28512b69014a2 WatchSource:0}: Error finding container ab924ef9bf311922a6ae1b8afad828fab05fec88ddd13d6a38d28512b69014a2: Status 404 returned error can't find the container with id ab924ef9bf311922a6ae1b8afad828fab05fec88ddd13d6a38d28512b69014a2 Dec 03 11:12:22 crc kubenswrapper[4646]: W1203 11:12:22.622563 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcec4fe45_3c8d_4dea_b5af_7fe6cff4594d.slice/crio-8617c24f4585e1ae3c3b2c2a6daadf9caab9ad17f3c9b7c34830dbe5685a1cf6 WatchSource:0}: Error finding container 8617c24f4585e1ae3c3b2c2a6daadf9caab9ad17f3c9b7c34830dbe5685a1cf6: Status 404 returned error can't find the container with id 8617c24f4585e1ae3c3b2c2a6daadf9caab9ad17f3c9b7c34830dbe5685a1cf6 Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.125049 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5129-account-create-update-drkql" event={"ID":"b5b8e5a9-ea00-4d58-8e49-50c72b9679de","Type":"ContainerStarted","Data":"0d7671ac9b7d09df4b158f59a928970acc197aae0d07f47ff0706f352fc827a6"} Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.125396 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5129-account-create-update-drkql" event={"ID":"b5b8e5a9-ea00-4d58-8e49-50c72b9679de","Type":"ContainerStarted","Data":"ab924ef9bf311922a6ae1b8afad828fab05fec88ddd13d6a38d28512b69014a2"} Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.130078 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tjvwv" event={"ID":"f23e454b-e13d-47f0-a3ba-2e96f3cc0649","Type":"ContainerStarted","Data":"f82c1610661677ce6cf8fa358dde61946710745257fd8990c4c0b9a5b1b869b5"} Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.132276 4646 generic.go:334] "Generic (PLEG): container finished" podID="16b24e6b-5a24-4abd-b3de-7932ce9e63c3" containerID="eb2590b56fd22c847e656d258a6487c8002c1fbefbf53a59191a36239cd17cbc" exitCode=0 Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.132366 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9dde-account-create-update-7ch8v" event={"ID":"16b24e6b-5a24-4abd-b3de-7932ce9e63c3","Type":"ContainerDied","Data":"eb2590b56fd22c847e656d258a6487c8002c1fbefbf53a59191a36239cd17cbc"} Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.132394 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9dde-account-create-update-7ch8v" event={"ID":"16b24e6b-5a24-4abd-b3de-7932ce9e63c3","Type":"ContainerStarted","Data":"f13a37fcec8f4e25176c042155a9a3d397c1d203d7007c7e0809768edfd3ef6e"} Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.134012 4646 generic.go:334] "Generic (PLEG): container finished" podID="76e53db6-897c-4c62-b732-034a119ab559" containerID="f3fcb07e0073c0900914442b0fd7bbbdfd572f596ff710351ab0ff11647202cb" exitCode=0 Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.134063 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-d95tn" event={"ID":"76e53db6-897c-4c62-b732-034a119ab559","Type":"ContainerDied","Data":"f3fcb07e0073c0900914442b0fd7bbbdfd572f596ff710351ab0ff11647202cb"} Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.134083 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-d95tn" event={"ID":"76e53db6-897c-4c62-b732-034a119ab559","Type":"ContainerStarted","Data":"e785fa97526b5c70b5e1af74a8198e3fce6bf9597a612799911405b9391e5955"} Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.135295 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jpkbq" event={"ID":"1785e265-f0f0-4e8f-8185-8f84b3121f3b","Type":"ContainerStarted","Data":"2f48166c546164da6a692b72a57229926c47d4b4a00683fdbac0c41a1491e15e"} Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.137085 4646 generic.go:334] "Generic (PLEG): container finished" podID="b26f36df-eade-45a7-84e5-a181b4b3d43f" containerID="8e45b35ae564b0b4890baacb3dd5dc19c5c7f59a2a9e10905e30881d50b17839" exitCode=0 Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.137141 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d2vp8" event={"ID":"b26f36df-eade-45a7-84e5-a181b4b3d43f","Type":"ContainerDied","Data":"8e45b35ae564b0b4890baacb3dd5dc19c5c7f59a2a9e10905e30881d50b17839"} Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.137160 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d2vp8" event={"ID":"b26f36df-eade-45a7-84e5-a181b4b3d43f","Type":"ContainerStarted","Data":"ecc17f38e541d7e7ae5fe93430da9482c78ccc94583d79e4d6be56db3467ab0f"} Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.146430 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-5129-account-create-update-drkql" podStartSLOduration=6.146413481 podStartE2EDuration="6.146413481s" podCreationTimestamp="2025-12-03 11:12:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:12:23.143276437 +0000 UTC m=+1119.606332592" watchObservedRunningTime="2025-12-03 11:12:23.146413481 +0000 UTC m=+1119.609469606" Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.158772 4646 generic.go:334] "Generic (PLEG): container finished" podID="cec4fe45-3c8d-4dea-b5af-7fe6cff4594d" containerID="20fcc2042ce42d7c3777f9b9634cfe0591710b4c0454a454d153e494e89fcb3d" exitCode=0 Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.158966 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-w557d" event={"ID":"cec4fe45-3c8d-4dea-b5af-7fe6cff4594d","Type":"ContainerDied","Data":"20fcc2042ce42d7c3777f9b9634cfe0591710b4c0454a454d153e494e89fcb3d"} Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.158992 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-w557d" event={"ID":"cec4fe45-3c8d-4dea-b5af-7fe6cff4594d","Type":"ContainerStarted","Data":"8617c24f4585e1ae3c3b2c2a6daadf9caab9ad17f3c9b7c34830dbe5685a1cf6"} Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.163776 4646 generic.go:334] "Generic (PLEG): container finished" podID="dc578e46-8ab2-4567-91ba-66462a04a328" containerID="bef107b3ef6f74e23fa5762c781ac1d36aaf0e7e4048eaa6ba8c0e0e64b9c3c7" exitCode=0 Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.163829 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e410-account-create-update-2nflx" event={"ID":"dc578e46-8ab2-4567-91ba-66462a04a328","Type":"ContainerDied","Data":"bef107b3ef6f74e23fa5762c781ac1d36aaf0e7e4048eaa6ba8c0e0e64b9c3c7"} Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.163859 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e410-account-create-update-2nflx" event={"ID":"dc578e46-8ab2-4567-91ba-66462a04a328","Type":"ContainerStarted","Data":"67f621298ca4faa316d821e9578986def8e82232eae9d8dc1f32033ce299f158"} Dec 03 11:12:23 crc kubenswrapper[4646]: I1203 11:12:23.178588 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-tjvwv" podStartSLOduration=2.857493857 podStartE2EDuration="18.17856266s" podCreationTimestamp="2025-12-03 11:12:05 +0000 UTC" firstStartedPulling="2025-12-03 11:12:06.704716302 +0000 UTC m=+1103.167772437" lastFinishedPulling="2025-12-03 11:12:22.025779765 +0000 UTC m=+1118.488841240" observedRunningTime="2025-12-03 11:12:23.17519803 +0000 UTC m=+1119.638254165" watchObservedRunningTime="2025-12-03 11:12:23.17856266 +0000 UTC m=+1119.641618785" Dec 03 11:12:24 crc kubenswrapper[4646]: I1203 11:12:24.175755 4646 generic.go:334] "Generic (PLEG): container finished" podID="b5b8e5a9-ea00-4d58-8e49-50c72b9679de" containerID="0d7671ac9b7d09df4b158f59a928970acc197aae0d07f47ff0706f352fc827a6" exitCode=0 Dec 03 11:12:24 crc kubenswrapper[4646]: I1203 11:12:24.175813 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5129-account-create-update-drkql" event={"ID":"b5b8e5a9-ea00-4d58-8e49-50c72b9679de","Type":"ContainerDied","Data":"0d7671ac9b7d09df4b158f59a928970acc197aae0d07f47ff0706f352fc827a6"} Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.200026 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-d95tn" event={"ID":"76e53db6-897c-4c62-b732-034a119ab559","Type":"ContainerDied","Data":"e785fa97526b5c70b5e1af74a8198e3fce6bf9597a612799911405b9391e5955"} Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.200328 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e785fa97526b5c70b5e1af74a8198e3fce6bf9597a612799911405b9391e5955" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.207045 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d2vp8" event={"ID":"b26f36df-eade-45a7-84e5-a181b4b3d43f","Type":"ContainerDied","Data":"ecc17f38e541d7e7ae5fe93430da9482c78ccc94583d79e4d6be56db3467ab0f"} Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.207308 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecc17f38e541d7e7ae5fe93430da9482c78ccc94583d79e4d6be56db3467ab0f" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.208831 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-w557d" event={"ID":"cec4fe45-3c8d-4dea-b5af-7fe6cff4594d","Type":"ContainerDied","Data":"8617c24f4585e1ae3c3b2c2a6daadf9caab9ad17f3c9b7c34830dbe5685a1cf6"} Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.209093 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8617c24f4585e1ae3c3b2c2a6daadf9caab9ad17f3c9b7c34830dbe5685a1cf6" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.210928 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e410-account-create-update-2nflx" event={"ID":"dc578e46-8ab2-4567-91ba-66462a04a328","Type":"ContainerDied","Data":"67f621298ca4faa316d821e9578986def8e82232eae9d8dc1f32033ce299f158"} Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.211087 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67f621298ca4faa316d821e9578986def8e82232eae9d8dc1f32033ce299f158" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.214133 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5129-account-create-update-drkql" event={"ID":"b5b8e5a9-ea00-4d58-8e49-50c72b9679de","Type":"ContainerDied","Data":"ab924ef9bf311922a6ae1b8afad828fab05fec88ddd13d6a38d28512b69014a2"} Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.214294 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab924ef9bf311922a6ae1b8afad828fab05fec88ddd13d6a38d28512b69014a2" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.216995 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9dde-account-create-update-7ch8v" event={"ID":"16b24e6b-5a24-4abd-b3de-7932ce9e63c3","Type":"ContainerDied","Data":"f13a37fcec8f4e25176c042155a9a3d397c1d203d7007c7e0809768edfd3ef6e"} Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.217146 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f13a37fcec8f4e25176c042155a9a3d397c1d203d7007c7e0809768edfd3ef6e" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.223247 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d2vp8" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.226322 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmxvt\" (UniqueName: \"kubernetes.io/projected/b26f36df-eade-45a7-84e5-a181b4b3d43f-kube-api-access-gmxvt\") pod \"b26f36df-eade-45a7-84e5-a181b4b3d43f\" (UID: \"b26f36df-eade-45a7-84e5-a181b4b3d43f\") " Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.226633 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b26f36df-eade-45a7-84e5-a181b4b3d43f-operator-scripts\") pod \"b26f36df-eade-45a7-84e5-a181b4b3d43f\" (UID: \"b26f36df-eade-45a7-84e5-a181b4b3d43f\") " Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.227504 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b26f36df-eade-45a7-84e5-a181b4b3d43f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b26f36df-eade-45a7-84e5-a181b4b3d43f" (UID: "b26f36df-eade-45a7-84e5-a181b4b3d43f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.233604 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b26f36df-eade-45a7-84e5-a181b4b3d43f-kube-api-access-gmxvt" (OuterVolumeSpecName: "kube-api-access-gmxvt") pod "b26f36df-eade-45a7-84e5-a181b4b3d43f" (UID: "b26f36df-eade-45a7-84e5-a181b4b3d43f"). InnerVolumeSpecName "kube-api-access-gmxvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.242554 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-w557d" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.313444 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9dde-account-create-update-7ch8v" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.329679 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b26f36df-eade-45a7-84e5-a181b4b3d43f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.329952 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmxvt\" (UniqueName: \"kubernetes.io/projected/b26f36df-eade-45a7-84e5-a181b4b3d43f-kube-api-access-gmxvt\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.331602 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e410-account-create-update-2nflx" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.357050 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-d95tn" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.364422 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5129-account-create-update-drkql" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.430822 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lt2r\" (UniqueName: \"kubernetes.io/projected/16b24e6b-5a24-4abd-b3de-7932ce9e63c3-kube-api-access-6lt2r\") pod \"16b24e6b-5a24-4abd-b3de-7932ce9e63c3\" (UID: \"16b24e6b-5a24-4abd-b3de-7932ce9e63c3\") " Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.430904 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16b24e6b-5a24-4abd-b3de-7932ce9e63c3-operator-scripts\") pod \"16b24e6b-5a24-4abd-b3de-7932ce9e63c3\" (UID: \"16b24e6b-5a24-4abd-b3de-7932ce9e63c3\") " Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.430941 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrw4t\" (UniqueName: \"kubernetes.io/projected/cec4fe45-3c8d-4dea-b5af-7fe6cff4594d-kube-api-access-rrw4t\") pod \"cec4fe45-3c8d-4dea-b5af-7fe6cff4594d\" (UID: \"cec4fe45-3c8d-4dea-b5af-7fe6cff4594d\") " Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.431008 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cec4fe45-3c8d-4dea-b5af-7fe6cff4594d-operator-scripts\") pod \"cec4fe45-3c8d-4dea-b5af-7fe6cff4594d\" (UID: \"cec4fe45-3c8d-4dea-b5af-7fe6cff4594d\") " Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.431547 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16b24e6b-5a24-4abd-b3de-7932ce9e63c3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "16b24e6b-5a24-4abd-b3de-7932ce9e63c3" (UID: "16b24e6b-5a24-4abd-b3de-7932ce9e63c3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.433034 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cec4fe45-3c8d-4dea-b5af-7fe6cff4594d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cec4fe45-3c8d-4dea-b5af-7fe6cff4594d" (UID: "cec4fe45-3c8d-4dea-b5af-7fe6cff4594d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.436175 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16b24e6b-5a24-4abd-b3de-7932ce9e63c3-kube-api-access-6lt2r" (OuterVolumeSpecName: "kube-api-access-6lt2r") pod "16b24e6b-5a24-4abd-b3de-7932ce9e63c3" (UID: "16b24e6b-5a24-4abd-b3de-7932ce9e63c3"). InnerVolumeSpecName "kube-api-access-6lt2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.438564 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cec4fe45-3c8d-4dea-b5af-7fe6cff4594d-kube-api-access-rrw4t" (OuterVolumeSpecName: "kube-api-access-rrw4t") pod "cec4fe45-3c8d-4dea-b5af-7fe6cff4594d" (UID: "cec4fe45-3c8d-4dea-b5af-7fe6cff4594d"). InnerVolumeSpecName "kube-api-access-rrw4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.532724 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jts9\" (UniqueName: \"kubernetes.io/projected/b5b8e5a9-ea00-4d58-8e49-50c72b9679de-kube-api-access-8jts9\") pod \"b5b8e5a9-ea00-4d58-8e49-50c72b9679de\" (UID: \"b5b8e5a9-ea00-4d58-8e49-50c72b9679de\") " Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.532777 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh6pj\" (UniqueName: \"kubernetes.io/projected/dc578e46-8ab2-4567-91ba-66462a04a328-kube-api-access-lh6pj\") pod \"dc578e46-8ab2-4567-91ba-66462a04a328\" (UID: \"dc578e46-8ab2-4567-91ba-66462a04a328\") " Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.532841 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc578e46-8ab2-4567-91ba-66462a04a328-operator-scripts\") pod \"dc578e46-8ab2-4567-91ba-66462a04a328\" (UID: \"dc578e46-8ab2-4567-91ba-66462a04a328\") " Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.532881 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76e53db6-897c-4c62-b732-034a119ab559-operator-scripts\") pod \"76e53db6-897c-4c62-b732-034a119ab559\" (UID: \"76e53db6-897c-4c62-b732-034a119ab559\") " Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.532919 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5b8e5a9-ea00-4d58-8e49-50c72b9679de-operator-scripts\") pod \"b5b8e5a9-ea00-4d58-8e49-50c72b9679de\" (UID: \"b5b8e5a9-ea00-4d58-8e49-50c72b9679de\") " Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.532968 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5kdf\" (UniqueName: \"kubernetes.io/projected/76e53db6-897c-4c62-b732-034a119ab559-kube-api-access-z5kdf\") pod \"76e53db6-897c-4c62-b732-034a119ab559\" (UID: \"76e53db6-897c-4c62-b732-034a119ab559\") " Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.533544 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76e53db6-897c-4c62-b732-034a119ab559-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "76e53db6-897c-4c62-b732-034a119ab559" (UID: "76e53db6-897c-4c62-b732-034a119ab559"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.533760 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5b8e5a9-ea00-4d58-8e49-50c72b9679de-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b5b8e5a9-ea00-4d58-8e49-50c72b9679de" (UID: "b5b8e5a9-ea00-4d58-8e49-50c72b9679de"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.533808 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc578e46-8ab2-4567-91ba-66462a04a328-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dc578e46-8ab2-4567-91ba-66462a04a328" (UID: "dc578e46-8ab2-4567-91ba-66462a04a328"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.534087 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrw4t\" (UniqueName: \"kubernetes.io/projected/cec4fe45-3c8d-4dea-b5af-7fe6cff4594d-kube-api-access-rrw4t\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.534106 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cec4fe45-3c8d-4dea-b5af-7fe6cff4594d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.534114 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc578e46-8ab2-4567-91ba-66462a04a328-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.534124 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76e53db6-897c-4c62-b732-034a119ab559-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.534133 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5b8e5a9-ea00-4d58-8e49-50c72b9679de-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.534142 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lt2r\" (UniqueName: \"kubernetes.io/projected/16b24e6b-5a24-4abd-b3de-7932ce9e63c3-kube-api-access-6lt2r\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.534151 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16b24e6b-5a24-4abd-b3de-7932ce9e63c3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.535969 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc578e46-8ab2-4567-91ba-66462a04a328-kube-api-access-lh6pj" (OuterVolumeSpecName: "kube-api-access-lh6pj") pod "dc578e46-8ab2-4567-91ba-66462a04a328" (UID: "dc578e46-8ab2-4567-91ba-66462a04a328"). InnerVolumeSpecName "kube-api-access-lh6pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.536992 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5b8e5a9-ea00-4d58-8e49-50c72b9679de-kube-api-access-8jts9" (OuterVolumeSpecName: "kube-api-access-8jts9") pod "b5b8e5a9-ea00-4d58-8e49-50c72b9679de" (UID: "b5b8e5a9-ea00-4d58-8e49-50c72b9679de"). InnerVolumeSpecName "kube-api-access-8jts9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.538128 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76e53db6-897c-4c62-b732-034a119ab559-kube-api-access-z5kdf" (OuterVolumeSpecName: "kube-api-access-z5kdf") pod "76e53db6-897c-4c62-b732-034a119ab559" (UID: "76e53db6-897c-4c62-b732-034a119ab559"). InnerVolumeSpecName "kube-api-access-z5kdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.635857 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jts9\" (UniqueName: \"kubernetes.io/projected/b5b8e5a9-ea00-4d58-8e49-50c72b9679de-kube-api-access-8jts9\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.635919 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh6pj\" (UniqueName: \"kubernetes.io/projected/dc578e46-8ab2-4567-91ba-66462a04a328-kube-api-access-lh6pj\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:27 crc kubenswrapper[4646]: I1203 11:12:27.635935 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5kdf\" (UniqueName: \"kubernetes.io/projected/76e53db6-897c-4c62-b732-034a119ab559-kube-api-access-z5kdf\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:28 crc kubenswrapper[4646]: I1203 11:12:28.229680 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e410-account-create-update-2nflx" Dec 03 11:12:28 crc kubenswrapper[4646]: I1203 11:12:28.230479 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9dde-account-create-update-7ch8v" Dec 03 11:12:28 crc kubenswrapper[4646]: I1203 11:12:28.231470 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-w557d" Dec 03 11:12:28 crc kubenswrapper[4646]: I1203 11:12:28.231889 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5129-account-create-update-drkql" Dec 03 11:12:28 crc kubenswrapper[4646]: I1203 11:12:28.232265 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-d95tn" Dec 03 11:12:28 crc kubenswrapper[4646]: I1203 11:12:28.232791 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d2vp8" Dec 03 11:12:28 crc kubenswrapper[4646]: I1203 11:12:28.233341 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jpkbq" event={"ID":"1785e265-f0f0-4e8f-8185-8f84b3121f3b","Type":"ContainerStarted","Data":"fbdb506e1582b61bbcd4f900d98a68fb89f18c9a3ac975b3cbe49dfec75563e2"} Dec 03 11:12:28 crc kubenswrapper[4646]: I1203 11:12:28.279179 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-jpkbq" podStartSLOduration=6.483783538 podStartE2EDuration="11.279149581s" podCreationTimestamp="2025-12-03 11:12:17 +0000 UTC" firstStartedPulling="2025-12-03 11:12:22.292222408 +0000 UTC m=+1118.755278543" lastFinishedPulling="2025-12-03 11:12:27.087588451 +0000 UTC m=+1123.550644586" observedRunningTime="2025-12-03 11:12:28.260722208 +0000 UTC m=+1124.723778343" watchObservedRunningTime="2025-12-03 11:12:28.279149581 +0000 UTC m=+1124.742205716" Dec 03 11:12:30 crc kubenswrapper[4646]: I1203 11:12:30.249202 4646 generic.go:334] "Generic (PLEG): container finished" podID="1785e265-f0f0-4e8f-8185-8f84b3121f3b" containerID="fbdb506e1582b61bbcd4f900d98a68fb89f18c9a3ac975b3cbe49dfec75563e2" exitCode=0 Dec 03 11:12:30 crc kubenswrapper[4646]: I1203 11:12:30.249291 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jpkbq" event={"ID":"1785e265-f0f0-4e8f-8185-8f84b3121f3b","Type":"ContainerDied","Data":"fbdb506e1582b61bbcd4f900d98a68fb89f18c9a3ac975b3cbe49dfec75563e2"} Dec 03 11:12:31 crc kubenswrapper[4646]: I1203 11:12:31.257856 4646 generic.go:334] "Generic (PLEG): container finished" podID="f23e454b-e13d-47f0-a3ba-2e96f3cc0649" containerID="f82c1610661677ce6cf8fa358dde61946710745257fd8990c4c0b9a5b1b869b5" exitCode=0 Dec 03 11:12:31 crc kubenswrapper[4646]: I1203 11:12:31.258084 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tjvwv" event={"ID":"f23e454b-e13d-47f0-a3ba-2e96f3cc0649","Type":"ContainerDied","Data":"f82c1610661677ce6cf8fa358dde61946710745257fd8990c4c0b9a5b1b869b5"} Dec 03 11:12:31 crc kubenswrapper[4646]: I1203 11:12:31.561775 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jpkbq" Dec 03 11:12:31 crc kubenswrapper[4646]: I1203 11:12:31.716560 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1785e265-f0f0-4e8f-8185-8f84b3121f3b-combined-ca-bundle\") pod \"1785e265-f0f0-4e8f-8185-8f84b3121f3b\" (UID: \"1785e265-f0f0-4e8f-8185-8f84b3121f3b\") " Dec 03 11:12:31 crc kubenswrapper[4646]: I1203 11:12:31.717583 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffx6l\" (UniqueName: \"kubernetes.io/projected/1785e265-f0f0-4e8f-8185-8f84b3121f3b-kube-api-access-ffx6l\") pod \"1785e265-f0f0-4e8f-8185-8f84b3121f3b\" (UID: \"1785e265-f0f0-4e8f-8185-8f84b3121f3b\") " Dec 03 11:12:31 crc kubenswrapper[4646]: I1203 11:12:31.718175 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1785e265-f0f0-4e8f-8185-8f84b3121f3b-config-data\") pod \"1785e265-f0f0-4e8f-8185-8f84b3121f3b\" (UID: \"1785e265-f0f0-4e8f-8185-8f84b3121f3b\") " Dec 03 11:12:31 crc kubenswrapper[4646]: I1203 11:12:31.721849 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1785e265-f0f0-4e8f-8185-8f84b3121f3b-kube-api-access-ffx6l" (OuterVolumeSpecName: "kube-api-access-ffx6l") pod "1785e265-f0f0-4e8f-8185-8f84b3121f3b" (UID: "1785e265-f0f0-4e8f-8185-8f84b3121f3b"). InnerVolumeSpecName "kube-api-access-ffx6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:31 crc kubenswrapper[4646]: I1203 11:12:31.741112 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1785e265-f0f0-4e8f-8185-8f84b3121f3b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1785e265-f0f0-4e8f-8185-8f84b3121f3b" (UID: "1785e265-f0f0-4e8f-8185-8f84b3121f3b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:12:31 crc kubenswrapper[4646]: I1203 11:12:31.764255 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1785e265-f0f0-4e8f-8185-8f84b3121f3b-config-data" (OuterVolumeSpecName: "config-data") pod "1785e265-f0f0-4e8f-8185-8f84b3121f3b" (UID: "1785e265-f0f0-4e8f-8185-8f84b3121f3b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:12:31 crc kubenswrapper[4646]: I1203 11:12:31.820039 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1785e265-f0f0-4e8f-8185-8f84b3121f3b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:31 crc kubenswrapper[4646]: I1203 11:12:31.820076 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffx6l\" (UniqueName: \"kubernetes.io/projected/1785e265-f0f0-4e8f-8185-8f84b3121f3b-kube-api-access-ffx6l\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:31 crc kubenswrapper[4646]: I1203 11:12:31.820093 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1785e265-f0f0-4e8f-8185-8f84b3121f3b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.274612 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jpkbq" event={"ID":"1785e265-f0f0-4e8f-8185-8f84b3121f3b","Type":"ContainerDied","Data":"2f48166c546164da6a692b72a57229926c47d4b4a00683fdbac0c41a1491e15e"} Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.274668 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f48166c546164da6a692b72a57229926c47d4b4a00683fdbac0c41a1491e15e" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.274634 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jpkbq" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.717638 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.843736 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-dvtgs"] Dec 03 11:12:32 crc kubenswrapper[4646]: E1203 11:12:32.844215 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b8e5a9-ea00-4d58-8e49-50c72b9679de" containerName="mariadb-account-create-update" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.844235 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b8e5a9-ea00-4d58-8e49-50c72b9679de" containerName="mariadb-account-create-update" Dec 03 11:12:32 crc kubenswrapper[4646]: E1203 11:12:32.844251 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc578e46-8ab2-4567-91ba-66462a04a328" containerName="mariadb-account-create-update" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.844257 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc578e46-8ab2-4567-91ba-66462a04a328" containerName="mariadb-account-create-update" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.844261 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpfz8\" (UniqueName: \"kubernetes.io/projected/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-kube-api-access-mpfz8\") pod \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " Dec 03 11:12:32 crc kubenswrapper[4646]: E1203 11:12:32.844275 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16b24e6b-5a24-4abd-b3de-7932ce9e63c3" containerName="mariadb-account-create-update" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.844411 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="16b24e6b-5a24-4abd-b3de-7932ce9e63c3" containerName="mariadb-account-create-update" Dec 03 11:12:32 crc kubenswrapper[4646]: E1203 11:12:32.844447 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76e53db6-897c-4c62-b732-034a119ab559" containerName="mariadb-database-create" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.844455 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="76e53db6-897c-4c62-b732-034a119ab559" containerName="mariadb-database-create" Dec 03 11:12:32 crc kubenswrapper[4646]: E1203 11:12:32.844467 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b26f36df-eade-45a7-84e5-a181b4b3d43f" containerName="mariadb-database-create" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.844473 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="b26f36df-eade-45a7-84e5-a181b4b3d43f" containerName="mariadb-database-create" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.844482 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-config-data\") pod \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.844521 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-db-sync-config-data\") pod \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.844565 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-combined-ca-bundle\") pod \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\" (UID: \"f23e454b-e13d-47f0-a3ba-2e96f3cc0649\") " Dec 03 11:12:32 crc kubenswrapper[4646]: E1203 11:12:32.844485 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1785e265-f0f0-4e8f-8185-8f84b3121f3b" containerName="keystone-db-sync" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.844637 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="1785e265-f0f0-4e8f-8185-8f84b3121f3b" containerName="keystone-db-sync" Dec 03 11:12:32 crc kubenswrapper[4646]: E1203 11:12:32.844717 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cec4fe45-3c8d-4dea-b5af-7fe6cff4594d" containerName="mariadb-database-create" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.844726 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="cec4fe45-3c8d-4dea-b5af-7fe6cff4594d" containerName="mariadb-database-create" Dec 03 11:12:32 crc kubenswrapper[4646]: E1203 11:12:32.844751 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f23e454b-e13d-47f0-a3ba-2e96f3cc0649" containerName="glance-db-sync" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.844757 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f23e454b-e13d-47f0-a3ba-2e96f3cc0649" containerName="glance-db-sync" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.845086 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="cec4fe45-3c8d-4dea-b5af-7fe6cff4594d" containerName="mariadb-database-create" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.845107 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc578e46-8ab2-4567-91ba-66462a04a328" containerName="mariadb-account-create-update" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.845124 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="1785e265-f0f0-4e8f-8185-8f84b3121f3b" containerName="keystone-db-sync" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.845136 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="f23e454b-e13d-47f0-a3ba-2e96f3cc0649" containerName="glance-db-sync" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.845146 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="16b24e6b-5a24-4abd-b3de-7932ce9e63c3" containerName="mariadb-account-create-update" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.845168 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="b26f36df-eade-45a7-84e5-a181b4b3d43f" containerName="mariadb-database-create" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.845178 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="76e53db6-897c-4c62-b732-034a119ab559" containerName="mariadb-database-create" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.845186 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5b8e5a9-ea00-4d58-8e49-50c72b9679de" containerName="mariadb-account-create-update" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.846296 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.848941 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f23e454b-e13d-47f0-a3ba-2e96f3cc0649" (UID: "f23e454b-e13d-47f0-a3ba-2e96f3cc0649"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.870965 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-kube-api-access-mpfz8" (OuterVolumeSpecName: "kube-api-access-mpfz8") pod "f23e454b-e13d-47f0-a3ba-2e96f3cc0649" (UID: "f23e454b-e13d-47f0-a3ba-2e96f3cc0649"). InnerVolumeSpecName "kube-api-access-mpfz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.878370 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-dvtgs"] Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.929509 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f23e454b-e13d-47f0-a3ba-2e96f3cc0649" (UID: "f23e454b-e13d-47f0-a3ba-2e96f3cc0649"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.946696 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-config-data" (OuterVolumeSpecName: "config-data") pod "f23e454b-e13d-47f0-a3ba-2e96f3cc0649" (UID: "f23e454b-e13d-47f0-a3ba-2e96f3cc0649"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.947658 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-config\") pod \"dnsmasq-dns-66fbd85b65-dvtgs\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.947734 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-dns-svc\") pod \"dnsmasq-dns-66fbd85b65-dvtgs\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.947834 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-ovsdbserver-sb\") pod \"dnsmasq-dns-66fbd85b65-dvtgs\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.947902 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-ovsdbserver-nb\") pod \"dnsmasq-dns-66fbd85b65-dvtgs\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.947950 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmx2q\" (UniqueName: \"kubernetes.io/projected/913b988e-bed2-4d69-90af-72d397584c9f-kube-api-access-xmx2q\") pod \"dnsmasq-dns-66fbd85b65-dvtgs\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.947994 4646 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.948016 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.948027 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpfz8\" (UniqueName: \"kubernetes.io/projected/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-kube-api-access-mpfz8\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.948037 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23e454b-e13d-47f0-a3ba-2e96f3cc0649-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.972252 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-jfdns"] Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.984283 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.989072 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.989237 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.989436 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.989528 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jfdns"] Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.989560 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 11:12:32 crc kubenswrapper[4646]: I1203 11:12:32.989684 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ddjxq" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.050261 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-ovsdbserver-nb\") pod \"dnsmasq-dns-66fbd85b65-dvtgs\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.050365 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmx2q\" (UniqueName: \"kubernetes.io/projected/913b988e-bed2-4d69-90af-72d397584c9f-kube-api-access-xmx2q\") pod \"dnsmasq-dns-66fbd85b65-dvtgs\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.050405 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-config\") pod \"dnsmasq-dns-66fbd85b65-dvtgs\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.050479 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-dns-svc\") pod \"dnsmasq-dns-66fbd85b65-dvtgs\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.050519 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-ovsdbserver-sb\") pod \"dnsmasq-dns-66fbd85b65-dvtgs\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.051447 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-ovsdbserver-nb\") pod \"dnsmasq-dns-66fbd85b65-dvtgs\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.051681 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-config\") pod \"dnsmasq-dns-66fbd85b65-dvtgs\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.051995 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-dns-svc\") pod \"dnsmasq-dns-66fbd85b65-dvtgs\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.052457 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-ovsdbserver-sb\") pod \"dnsmasq-dns-66fbd85b65-dvtgs\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.094520 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmx2q\" (UniqueName: \"kubernetes.io/projected/913b988e-bed2-4d69-90af-72d397584c9f-kube-api-access-xmx2q\") pod \"dnsmasq-dns-66fbd85b65-dvtgs\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.152492 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-fernet-keys\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.152548 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-credential-keys\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.152577 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4x8g\" (UniqueName: \"kubernetes.io/projected/f706e096-2b8c-4696-a52f-a2e200508170-kube-api-access-r4x8g\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.152680 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-combined-ca-bundle\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.152710 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-scripts\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.152764 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-config-data\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.219522 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.221431 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.254450 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-config-data\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.254543 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-fernet-keys\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.254579 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-credential-keys\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.254601 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4x8g\" (UniqueName: \"kubernetes.io/projected/f706e096-2b8c-4696-a52f-a2e200508170-kube-api-access-r4x8g\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.254638 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-combined-ca-bundle\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.254654 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-scripts\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.265005 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.266116 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.270689 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-scripts\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.271017 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-credential-keys\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.274076 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-combined-ca-bundle\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.274093 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-config-data\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.284692 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tjvwv" event={"ID":"f23e454b-e13d-47f0-a3ba-2e96f3cc0649","Type":"ContainerDied","Data":"0322f616c0e6df2162e21879f2b3bd99c1066db369fabfda898d47f01f9b2c6d"} Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.284734 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0322f616c0e6df2162e21879f2b3bd99c1066db369fabfda898d47f01f9b2c6d" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.284790 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tjvwv" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.287592 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-2nh9f"] Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.288619 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2nh9f" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.290327 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-fernet-keys\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.300797 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.301004 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-2zxxm" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.301123 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.303032 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.310866 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.322541 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-rtwcp"] Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.324151 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.327529 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.328925 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-cvt9x" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.329246 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4x8g\" (UniqueName: \"kubernetes.io/projected/f706e096-2b8c-4696-a52f-a2e200508170-kube-api-access-r4x8g\") pod \"keystone-bootstrap-jfdns\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.335126 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2nh9f"] Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.335603 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.343925 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-h4hq8"] Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.345288 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h4hq8" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.363141 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-x2m59" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.363618 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.366167 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7715c927-8035-48d3-908e-4f3838f296aa-run-httpd\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.366375 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.366485 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7715c927-8035-48d3-908e-4f3838f296aa-log-httpd\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.366654 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-config-data\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.366789 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cx9n\" (UniqueName: \"kubernetes.io/projected/7715c927-8035-48d3-908e-4f3838f296aa-kube-api-access-2cx9n\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.366819 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.366949 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-scripts\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.399971 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-rtwcp"] Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.476183 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-combined-ca-bundle\") pod \"barbican-db-sync-h4hq8\" (UID: \"411d5bc2-b69a-4b93-bb7e-c8990b44e79a\") " pod="openstack/barbican-db-sync-h4hq8" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.476241 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh5xk\" (UniqueName: \"kubernetes.io/projected/aad954e4-5d92-4b2b-9771-a5f78f11d169-kube-api-access-lh5xk\") pod \"neutron-db-sync-2nh9f\" (UID: \"aad954e4-5d92-4b2b-9771-a5f78f11d169\") " pod="openstack/neutron-db-sync-2nh9f" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.476259 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-db-sync-config-data\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.476279 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad954e4-5d92-4b2b-9771-a5f78f11d169-combined-ca-bundle\") pod \"neutron-db-sync-2nh9f\" (UID: \"aad954e4-5d92-4b2b-9771-a5f78f11d169\") " pod="openstack/neutron-db-sync-2nh9f" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.476315 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7715c927-8035-48d3-908e-4f3838f296aa-run-httpd\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.481391 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-h4hq8"] Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.493481 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-db-sync-config-data\") pod \"barbican-db-sync-h4hq8\" (UID: \"411d5bc2-b69a-4b93-bb7e-c8990b44e79a\") " pod="openstack/barbican-db-sync-h4hq8" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.493549 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-config-data\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.493577 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.493622 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6rd9\" (UniqueName: \"kubernetes.io/projected/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-kube-api-access-f6rd9\") pod \"barbican-db-sync-h4hq8\" (UID: \"411d5bc2-b69a-4b93-bb7e-c8990b44e79a\") " pod="openstack/barbican-db-sync-h4hq8" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.493661 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7715c927-8035-48d3-908e-4f3838f296aa-log-httpd\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.493685 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76443d6c-7930-4c62-abbb-d7ff831232ca-etc-machine-id\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.493705 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-combined-ca-bundle\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.493727 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-scripts\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.493769 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-config-data\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.493806 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cx9n\" (UniqueName: \"kubernetes.io/projected/7715c927-8035-48d3-908e-4f3838f296aa-kube-api-access-2cx9n\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.493823 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.493857 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzqhk\" (UniqueName: \"kubernetes.io/projected/76443d6c-7930-4c62-abbb-d7ff831232ca-kube-api-access-vzqhk\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.493891 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-scripts\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.493935 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aad954e4-5d92-4b2b-9771-a5f78f11d169-config\") pod \"neutron-db-sync-2nh9f\" (UID: \"aad954e4-5d92-4b2b-9771-a5f78f11d169\") " pod="openstack/neutron-db-sync-2nh9f" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.494388 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7715c927-8035-48d3-908e-4f3838f296aa-run-httpd\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.499791 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7715c927-8035-48d3-908e-4f3838f296aa-log-httpd\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.500607 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-scripts\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.501938 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.504975 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-config-data\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.515145 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.519728 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-dvtgs"] Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.531286 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-ghphj"] Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.532431 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.540659 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-g4pdh" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.540856 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.540984 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.541834 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-7565b"] Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.542110 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cx9n\" (UniqueName: \"kubernetes.io/projected/7715c927-8035-48d3-908e-4f3838f296aa-kube-api-access-2cx9n\") pod \"ceilometer-0\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.543361 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.555037 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ghphj"] Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.568261 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-7565b"] Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.595032 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzqhk\" (UniqueName: \"kubernetes.io/projected/76443d6c-7930-4c62-abbb-d7ff831232ca-kube-api-access-vzqhk\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.595086 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aad954e4-5d92-4b2b-9771-a5f78f11d169-config\") pod \"neutron-db-sync-2nh9f\" (UID: \"aad954e4-5d92-4b2b-9771-a5f78f11d169\") " pod="openstack/neutron-db-sync-2nh9f" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.595118 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-combined-ca-bundle\") pod \"barbican-db-sync-h4hq8\" (UID: \"411d5bc2-b69a-4b93-bb7e-c8990b44e79a\") " pod="openstack/barbican-db-sync-h4hq8" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.595143 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh5xk\" (UniqueName: \"kubernetes.io/projected/aad954e4-5d92-4b2b-9771-a5f78f11d169-kube-api-access-lh5xk\") pod \"neutron-db-sync-2nh9f\" (UID: \"aad954e4-5d92-4b2b-9771-a5f78f11d169\") " pod="openstack/neutron-db-sync-2nh9f" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.595159 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-db-sync-config-data\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.595178 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad954e4-5d92-4b2b-9771-a5f78f11d169-combined-ca-bundle\") pod \"neutron-db-sync-2nh9f\" (UID: \"aad954e4-5d92-4b2b-9771-a5f78f11d169\") " pod="openstack/neutron-db-sync-2nh9f" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.595212 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-db-sync-config-data\") pod \"barbican-db-sync-h4hq8\" (UID: \"411d5bc2-b69a-4b93-bb7e-c8990b44e79a\") " pod="openstack/barbican-db-sync-h4hq8" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.595235 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-config-data\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.595258 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6rd9\" (UniqueName: \"kubernetes.io/projected/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-kube-api-access-f6rd9\") pod \"barbican-db-sync-h4hq8\" (UID: \"411d5bc2-b69a-4b93-bb7e-c8990b44e79a\") " pod="openstack/barbican-db-sync-h4hq8" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.595283 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76443d6c-7930-4c62-abbb-d7ff831232ca-etc-machine-id\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.595298 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-combined-ca-bundle\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.595314 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-scripts\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.597937 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76443d6c-7930-4c62-abbb-d7ff831232ca-etc-machine-id\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.612553 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-combined-ca-bundle\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.627218 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-db-sync-config-data\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.628702 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.629174 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad954e4-5d92-4b2b-9771-a5f78f11d169-combined-ca-bundle\") pod \"neutron-db-sync-2nh9f\" (UID: \"aad954e4-5d92-4b2b-9771-a5f78f11d169\") " pod="openstack/neutron-db-sync-2nh9f" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.630587 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-db-sync-config-data\") pod \"barbican-db-sync-h4hq8\" (UID: \"411d5bc2-b69a-4b93-bb7e-c8990b44e79a\") " pod="openstack/barbican-db-sync-h4hq8" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.633842 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-combined-ca-bundle\") pod \"barbican-db-sync-h4hq8\" (UID: \"411d5bc2-b69a-4b93-bb7e-c8990b44e79a\") " pod="openstack/barbican-db-sync-h4hq8" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.637048 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzqhk\" (UniqueName: \"kubernetes.io/projected/76443d6c-7930-4c62-abbb-d7ff831232ca-kube-api-access-vzqhk\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.639372 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-scripts\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.641436 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-config-data\") pod \"cinder-db-sync-rtwcp\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.647077 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh5xk\" (UniqueName: \"kubernetes.io/projected/aad954e4-5d92-4b2b-9771-a5f78f11d169-kube-api-access-lh5xk\") pod \"neutron-db-sync-2nh9f\" (UID: \"aad954e4-5d92-4b2b-9771-a5f78f11d169\") " pod="openstack/neutron-db-sync-2nh9f" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.648619 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/aad954e4-5d92-4b2b-9771-a5f78f11d169-config\") pod \"neutron-db-sync-2nh9f\" (UID: \"aad954e4-5d92-4b2b-9771-a5f78f11d169\") " pod="openstack/neutron-db-sync-2nh9f" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.700636 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf59f66bf-7565b\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.700857 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-combined-ca-bundle\") pod \"placement-db-sync-ghphj\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.700976 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-dns-svc\") pod \"dnsmasq-dns-6bf59f66bf-7565b\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.701077 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-scripts\") pod \"placement-db-sync-ghphj\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.701177 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0028dab-487a-4f33-be6b-0f082d4f617e-logs\") pod \"placement-db-sync-ghphj\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.701292 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs4z9\" (UniqueName: \"kubernetes.io/projected/e8e61d88-5b29-4886-80aa-3a818c0594c6-kube-api-access-bs4z9\") pod \"dnsmasq-dns-6bf59f66bf-7565b\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.701399 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-config\") pod \"dnsmasq-dns-6bf59f66bf-7565b\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.701477 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-config-data\") pod \"placement-db-sync-ghphj\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.701554 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf59f66bf-7565b\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.701713 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trtqv\" (UniqueName: \"kubernetes.io/projected/c0028dab-487a-4f33-be6b-0f082d4f617e-kube-api-access-trtqv\") pod \"placement-db-sync-ghphj\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.702403 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6rd9\" (UniqueName: \"kubernetes.io/projected/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-kube-api-access-f6rd9\") pod \"barbican-db-sync-h4hq8\" (UID: \"411d5bc2-b69a-4b93-bb7e-c8990b44e79a\") " pod="openstack/barbican-db-sync-h4hq8" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.707166 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.804154 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0028dab-487a-4f33-be6b-0f082d4f617e-logs\") pod \"placement-db-sync-ghphj\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.804218 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs4z9\" (UniqueName: \"kubernetes.io/projected/e8e61d88-5b29-4886-80aa-3a818c0594c6-kube-api-access-bs4z9\") pod \"dnsmasq-dns-6bf59f66bf-7565b\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.804240 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-config\") pod \"dnsmasq-dns-6bf59f66bf-7565b\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.804256 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-config-data\") pod \"placement-db-sync-ghphj\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.804646 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0028dab-487a-4f33-be6b-0f082d4f617e-logs\") pod \"placement-db-sync-ghphj\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.804692 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf59f66bf-7565b\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.804709 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trtqv\" (UniqueName: \"kubernetes.io/projected/c0028dab-487a-4f33-be6b-0f082d4f617e-kube-api-access-trtqv\") pod \"placement-db-sync-ghphj\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.805218 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf59f66bf-7565b\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.805237 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-combined-ca-bundle\") pod \"placement-db-sync-ghphj\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.805269 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-dns-svc\") pod \"dnsmasq-dns-6bf59f66bf-7565b\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.805290 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-scripts\") pod \"placement-db-sync-ghphj\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.805406 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-config\") pod \"dnsmasq-dns-6bf59f66bf-7565b\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.805737 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf59f66bf-7565b\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.806243 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-dns-svc\") pod \"dnsmasq-dns-6bf59f66bf-7565b\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.806279 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf59f66bf-7565b\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.816986 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-config-data\") pod \"placement-db-sync-ghphj\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.846986 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-combined-ca-bundle\") pod \"placement-db-sync-ghphj\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.847000 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.892550 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-scripts\") pod \"placement-db-sync-ghphj\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.893047 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trtqv\" (UniqueName: \"kubernetes.io/projected/c0028dab-487a-4f33-be6b-0f082d4f617e-kube-api-access-trtqv\") pod \"placement-db-sync-ghphj\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.926377 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs4z9\" (UniqueName: \"kubernetes.io/projected/e8e61d88-5b29-4886-80aa-3a818c0594c6-kube-api-access-bs4z9\") pod \"dnsmasq-dns-6bf59f66bf-7565b\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:33 crc kubenswrapper[4646]: I1203 11:12:33.956778 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2nh9f" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.003808 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h4hq8" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.026936 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-7565b"] Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.027854 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.070025 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf"] Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.079911 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.105749 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf"] Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.170247 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ghphj" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.227083 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5m7w\" (UniqueName: \"kubernetes.io/projected/1495a64c-4fe7-4875-ae31-05ac0f3949b8-kube-api-access-h5m7w\") pod \"dnsmasq-dns-5b6dbdb6f5-5wrhf\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.227123 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-dns-svc\") pod \"dnsmasq-dns-5b6dbdb6f5-5wrhf\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.227157 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6dbdb6f5-5wrhf\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.227227 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-config\") pod \"dnsmasq-dns-5b6dbdb6f5-5wrhf\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.227251 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-ovsdbserver-sb\") pod \"dnsmasq-dns-5b6dbdb6f5-5wrhf\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.318925 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-dvtgs"] Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.328767 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5m7w\" (UniqueName: \"kubernetes.io/projected/1495a64c-4fe7-4875-ae31-05ac0f3949b8-kube-api-access-h5m7w\") pod \"dnsmasq-dns-5b6dbdb6f5-5wrhf\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.328801 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-dns-svc\") pod \"dnsmasq-dns-5b6dbdb6f5-5wrhf\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.328836 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6dbdb6f5-5wrhf\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.328902 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-config\") pod \"dnsmasq-dns-5b6dbdb6f5-5wrhf\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.328928 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-ovsdbserver-sb\") pod \"dnsmasq-dns-5b6dbdb6f5-5wrhf\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.329563 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6dbdb6f5-5wrhf\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.329581 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-ovsdbserver-sb\") pod \"dnsmasq-dns-5b6dbdb6f5-5wrhf\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.330059 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-dns-svc\") pod \"dnsmasq-dns-5b6dbdb6f5-5wrhf\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.330897 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-config\") pod \"dnsmasq-dns-5b6dbdb6f5-5wrhf\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.373652 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5m7w\" (UniqueName: \"kubernetes.io/projected/1495a64c-4fe7-4875-ae31-05ac0f3949b8-kube-api-access-h5m7w\") pod \"dnsmasq-dns-5b6dbdb6f5-5wrhf\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.424172 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.581832 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-rtwcp"] Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.836749 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jfdns"] Dec 03 11:12:34 crc kubenswrapper[4646]: W1203 11:12:34.858581 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf706e096_2b8c_4696_a52f_a2e200508170.slice/crio-3b4906c367df01a988d432c924c22ff85f547f80da27e542df7d8c374da140ff WatchSource:0}: Error finding container 3b4906c367df01a988d432c924c22ff85f547f80da27e542df7d8c374da140ff: Status 404 returned error can't find the container with id 3b4906c367df01a988d432c924c22ff85f547f80da27e542df7d8c374da140ff Dec 03 11:12:34 crc kubenswrapper[4646]: I1203 11:12:34.913241 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-7565b"] Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.036430 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2nh9f"] Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.046289 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.054402 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ghphj"] Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.065270 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-h4hq8"] Dec 03 11:12:35 crc kubenswrapper[4646]: W1203 11:12:35.094357 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0028dab_487a_4f33_be6b_0f082d4f617e.slice/crio-c55e11fa607664f234091d33313c8cf6bd4c6f8a4d93458e7736775908083555 WatchSource:0}: Error finding container c55e11fa607664f234091d33313c8cf6bd4c6f8a4d93458e7736775908083555: Status 404 returned error can't find the container with id c55e11fa607664f234091d33313c8cf6bd4c6f8a4d93458e7736775908083555 Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.267170 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf"] Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.321902 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rtwcp" event={"ID":"76443d6c-7930-4c62-abbb-d7ff831232ca","Type":"ContainerStarted","Data":"3b715aa20bee782c924fe115d7211eb838580ee2cee80a46adc742debcc50616"} Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.324476 4646 generic.go:334] "Generic (PLEG): container finished" podID="e8e61d88-5b29-4886-80aa-3a818c0594c6" containerID="0ba1026cd134f2c31855b1f937794c3bf774404e445fdfe2864f7f1f5836c2d6" exitCode=0 Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.324533 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" event={"ID":"e8e61d88-5b29-4886-80aa-3a818c0594c6","Type":"ContainerDied","Data":"0ba1026cd134f2c31855b1f937794c3bf774404e445fdfe2864f7f1f5836c2d6"} Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.324558 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" event={"ID":"e8e61d88-5b29-4886-80aa-3a818c0594c6","Type":"ContainerStarted","Data":"1827647a521e17808a97aa81e5d01871686164048d55eebeb551b7bb1c158893"} Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.328445 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2nh9f" event={"ID":"aad954e4-5d92-4b2b-9771-a5f78f11d169","Type":"ContainerStarted","Data":"e1b8411e617482fad0b1b766802894be967bdb8788000ca6e634171ff64f7bb1"} Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.329716 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h4hq8" event={"ID":"411d5bc2-b69a-4b93-bb7e-c8990b44e79a","Type":"ContainerStarted","Data":"2fb16059442c4c64e7e78a99fa95bfc4ad1e8e5472d3f0587adaa99dd2a83017"} Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.330875 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7715c927-8035-48d3-908e-4f3838f296aa","Type":"ContainerStarted","Data":"6da83e9f1049a4caafa2d3961db0d5afb1c058ebdfd632ce1b7f72842ce5a704"} Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.334147 4646 generic.go:334] "Generic (PLEG): container finished" podID="913b988e-bed2-4d69-90af-72d397584c9f" containerID="8b59095b6df486f448a72c405e7680e9f087b039c1db8b989a0877a7a1b018d9" exitCode=0 Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.334237 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" event={"ID":"913b988e-bed2-4d69-90af-72d397584c9f","Type":"ContainerDied","Data":"8b59095b6df486f448a72c405e7680e9f087b039c1db8b989a0877a7a1b018d9"} Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.334385 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" event={"ID":"913b988e-bed2-4d69-90af-72d397584c9f","Type":"ContainerStarted","Data":"591a89a47cad9feee4c0881f1d58772b42f36cac10673f5c72e32f9e1b2c28fc"} Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.338121 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ghphj" event={"ID":"c0028dab-487a-4f33-be6b-0f082d4f617e","Type":"ContainerStarted","Data":"c55e11fa607664f234091d33313c8cf6bd4c6f8a4d93458e7736775908083555"} Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.343548 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jfdns" event={"ID":"f706e096-2b8c-4696-a52f-a2e200508170","Type":"ContainerStarted","Data":"6462075ca2426030733798995ebbc6119432ea1d42417a8a33ba051445c2bda9"} Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.343633 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jfdns" event={"ID":"f706e096-2b8c-4696-a52f-a2e200508170","Type":"ContainerStarted","Data":"3b4906c367df01a988d432c924c22ff85f547f80da27e542df7d8c374da140ff"} Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.381847 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-jfdns" podStartSLOduration=3.381823399 podStartE2EDuration="3.381823399s" podCreationTimestamp="2025-12-03 11:12:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:12:35.377640327 +0000 UTC m=+1131.840696462" watchObservedRunningTime="2025-12-03 11:12:35.381823399 +0000 UTC m=+1131.844879544" Dec 03 11:12:35 crc kubenswrapper[4646]: I1203 11:12:35.885378 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.011724 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.085583 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-ovsdbserver-nb\") pod \"e8e61d88-5b29-4886-80aa-3a818c0594c6\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.085632 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs4z9\" (UniqueName: \"kubernetes.io/projected/e8e61d88-5b29-4886-80aa-3a818c0594c6-kube-api-access-bs4z9\") pod \"e8e61d88-5b29-4886-80aa-3a818c0594c6\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.087790 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-dns-svc\") pod \"e8e61d88-5b29-4886-80aa-3a818c0594c6\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.087864 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-config\") pod \"e8e61d88-5b29-4886-80aa-3a818c0594c6\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.087884 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-ovsdbserver-sb\") pod \"e8e61d88-5b29-4886-80aa-3a818c0594c6\" (UID: \"e8e61d88-5b29-4886-80aa-3a818c0594c6\") " Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.105601 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8e61d88-5b29-4886-80aa-3a818c0594c6-kube-api-access-bs4z9" (OuterVolumeSpecName: "kube-api-access-bs4z9") pod "e8e61d88-5b29-4886-80aa-3a818c0594c6" (UID: "e8e61d88-5b29-4886-80aa-3a818c0594c6"). InnerVolumeSpecName "kube-api-access-bs4z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.115426 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e8e61d88-5b29-4886-80aa-3a818c0594c6" (UID: "e8e61d88-5b29-4886-80aa-3a818c0594c6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.116432 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-config" (OuterVolumeSpecName: "config") pod "e8e61d88-5b29-4886-80aa-3a818c0594c6" (UID: "e8e61d88-5b29-4886-80aa-3a818c0594c6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.121868 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e8e61d88-5b29-4886-80aa-3a818c0594c6" (UID: "e8e61d88-5b29-4886-80aa-3a818c0594c6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.153738 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e8e61d88-5b29-4886-80aa-3a818c0594c6" (UID: "e8e61d88-5b29-4886-80aa-3a818c0594c6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.189318 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-config\") pod \"913b988e-bed2-4d69-90af-72d397584c9f\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.189713 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmx2q\" (UniqueName: \"kubernetes.io/projected/913b988e-bed2-4d69-90af-72d397584c9f-kube-api-access-xmx2q\") pod \"913b988e-bed2-4d69-90af-72d397584c9f\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.189751 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-dns-svc\") pod \"913b988e-bed2-4d69-90af-72d397584c9f\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.189786 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-ovsdbserver-sb\") pod \"913b988e-bed2-4d69-90af-72d397584c9f\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.189863 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-ovsdbserver-nb\") pod \"913b988e-bed2-4d69-90af-72d397584c9f\" (UID: \"913b988e-bed2-4d69-90af-72d397584c9f\") " Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.190672 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.190692 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.190703 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.190714 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs4z9\" (UniqueName: \"kubernetes.io/projected/e8e61d88-5b29-4886-80aa-3a818c0594c6-kube-api-access-bs4z9\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.190724 4646 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8e61d88-5b29-4886-80aa-3a818c0594c6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.198189 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/913b988e-bed2-4d69-90af-72d397584c9f-kube-api-access-xmx2q" (OuterVolumeSpecName: "kube-api-access-xmx2q") pod "913b988e-bed2-4d69-90af-72d397584c9f" (UID: "913b988e-bed2-4d69-90af-72d397584c9f"). InnerVolumeSpecName "kube-api-access-xmx2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.231064 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-config" (OuterVolumeSpecName: "config") pod "913b988e-bed2-4d69-90af-72d397584c9f" (UID: "913b988e-bed2-4d69-90af-72d397584c9f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.232494 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "913b988e-bed2-4d69-90af-72d397584c9f" (UID: "913b988e-bed2-4d69-90af-72d397584c9f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.234921 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "913b988e-bed2-4d69-90af-72d397584c9f" (UID: "913b988e-bed2-4d69-90af-72d397584c9f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.239223 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "913b988e-bed2-4d69-90af-72d397584c9f" (UID: "913b988e-bed2-4d69-90af-72d397584c9f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.295636 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.295945 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmx2q\" (UniqueName: \"kubernetes.io/projected/913b988e-bed2-4d69-90af-72d397584c9f-kube-api-access-xmx2q\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.295958 4646 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.295969 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.296560 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/913b988e-bed2-4d69-90af-72d397584c9f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.356999 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.359834 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf59f66bf-7565b" event={"ID":"e8e61d88-5b29-4886-80aa-3a818c0594c6","Type":"ContainerDied","Data":"1827647a521e17808a97aa81e5d01871686164048d55eebeb551b7bb1c158893"} Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.359891 4646 scope.go:117] "RemoveContainer" containerID="0ba1026cd134f2c31855b1f937794c3bf774404e445fdfe2864f7f1f5836c2d6" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.368651 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2nh9f" event={"ID":"aad954e4-5d92-4b2b-9771-a5f78f11d169","Type":"ContainerStarted","Data":"0b73033a13582c0cbbc6b90aa28c4838f5894f967db61e4e58282e22bebaab17"} Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.375141 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" event={"ID":"913b988e-bed2-4d69-90af-72d397584c9f","Type":"ContainerDied","Data":"591a89a47cad9feee4c0881f1d58772b42f36cac10673f5c72e32f9e1b2c28fc"} Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.375418 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fbd85b65-dvtgs" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.405947 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-2nh9f" podStartSLOduration=3.405928464 podStartE2EDuration="3.405928464s" podCreationTimestamp="2025-12-03 11:12:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:12:36.389868145 +0000 UTC m=+1132.852924280" watchObservedRunningTime="2025-12-03 11:12:36.405928464 +0000 UTC m=+1132.868984599" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.409447 4646 generic.go:334] "Generic (PLEG): container finished" podID="1495a64c-4fe7-4875-ae31-05ac0f3949b8" containerID="a28909ef1d585195d0d1e36582d8c3ebf0c1cc15e627c2fefaae4e77a6f6c647" exitCode=0 Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.409687 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" event={"ID":"1495a64c-4fe7-4875-ae31-05ac0f3949b8","Type":"ContainerDied","Data":"a28909ef1d585195d0d1e36582d8c3ebf0c1cc15e627c2fefaae4e77a6f6c647"} Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.409841 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" event={"ID":"1495a64c-4fe7-4875-ae31-05ac0f3949b8","Type":"ContainerStarted","Data":"6cd550227fb32b9ba35b80943e871c1e7bc303b1dc2a4666e6f555f80ae49dd0"} Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.455089 4646 scope.go:117] "RemoveContainer" containerID="8b59095b6df486f448a72c405e7680e9f087b039c1db8b989a0877a7a1b018d9" Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.502852 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-7565b"] Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.523407 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-7565b"] Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.586432 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-dvtgs"] Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.599928 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-dvtgs"] Dec 03 11:12:36 crc kubenswrapper[4646]: I1203 11:12:36.788432 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:12:37 crc kubenswrapper[4646]: I1203 11:12:37.443078 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" event={"ID":"1495a64c-4fe7-4875-ae31-05ac0f3949b8","Type":"ContainerStarted","Data":"91a79db0696da530f2cc08a85addb1b08293ef5ce79f580b8814f87f063caa45"} Dec 03 11:12:37 crc kubenswrapper[4646]: I1203 11:12:37.443287 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:37 crc kubenswrapper[4646]: I1203 11:12:37.472474 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" podStartSLOduration=3.472457304 podStartE2EDuration="3.472457304s" podCreationTimestamp="2025-12-03 11:12:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:12:37.464262064 +0000 UTC m=+1133.927318229" watchObservedRunningTime="2025-12-03 11:12:37.472457304 +0000 UTC m=+1133.935513439" Dec 03 11:12:37 crc kubenswrapper[4646]: I1203 11:12:37.864183 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="913b988e-bed2-4d69-90af-72d397584c9f" path="/var/lib/kubelet/pods/913b988e-bed2-4d69-90af-72d397584c9f/volumes" Dec 03 11:12:37 crc kubenswrapper[4646]: I1203 11:12:37.864763 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8e61d88-5b29-4886-80aa-3a818c0594c6" path="/var/lib/kubelet/pods/e8e61d88-5b29-4886-80aa-3a818c0594c6/volumes" Dec 03 11:12:41 crc kubenswrapper[4646]: I1203 11:12:41.500705 4646 generic.go:334] "Generic (PLEG): container finished" podID="f706e096-2b8c-4696-a52f-a2e200508170" containerID="6462075ca2426030733798995ebbc6119432ea1d42417a8a33ba051445c2bda9" exitCode=0 Dec 03 11:12:41 crc kubenswrapper[4646]: I1203 11:12:41.501059 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jfdns" event={"ID":"f706e096-2b8c-4696-a52f-a2e200508170","Type":"ContainerDied","Data":"6462075ca2426030733798995ebbc6119432ea1d42417a8a33ba051445c2bda9"} Dec 03 11:12:44 crc kubenswrapper[4646]: I1203 11:12:44.431451 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:12:44 crc kubenswrapper[4646]: I1203 11:12:44.497669 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-tqqfs"] Dec 03 11:12:44 crc kubenswrapper[4646]: I1203 11:12:44.497882 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-tqqfs" podUID="758d0aa0-8428-46bb-902d-f751b670c1ac" containerName="dnsmasq-dns" containerID="cri-o://8adc4d7cca6da85aa4a5b3e75931c2edb7f7d3204c30b6e47348f912816cb87f" gracePeriod=10 Dec 03 11:12:45 crc kubenswrapper[4646]: I1203 11:12:45.573909 4646 generic.go:334] "Generic (PLEG): container finished" podID="758d0aa0-8428-46bb-902d-f751b670c1ac" containerID="8adc4d7cca6da85aa4a5b3e75931c2edb7f7d3204c30b6e47348f912816cb87f" exitCode=0 Dec 03 11:12:45 crc kubenswrapper[4646]: I1203 11:12:45.574038 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-tqqfs" event={"ID":"758d0aa0-8428-46bb-902d-f751b670c1ac","Type":"ContainerDied","Data":"8adc4d7cca6da85aa4a5b3e75931c2edb7f7d3204c30b6e47348f912816cb87f"} Dec 03 11:12:47 crc kubenswrapper[4646]: I1203 11:12:47.658934 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-tqqfs" podUID="758d0aa0-8428-46bb-902d-f751b670c1ac" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: connect: connection refused" Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.283195 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.329733 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-credential-keys\") pod \"f706e096-2b8c-4696-a52f-a2e200508170\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.329789 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-fernet-keys\") pod \"f706e096-2b8c-4696-a52f-a2e200508170\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.329829 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-config-data\") pod \"f706e096-2b8c-4696-a52f-a2e200508170\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.329851 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4x8g\" (UniqueName: \"kubernetes.io/projected/f706e096-2b8c-4696-a52f-a2e200508170-kube-api-access-r4x8g\") pod \"f706e096-2b8c-4696-a52f-a2e200508170\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.329928 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-combined-ca-bundle\") pod \"f706e096-2b8c-4696-a52f-a2e200508170\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.329991 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-scripts\") pod \"f706e096-2b8c-4696-a52f-a2e200508170\" (UID: \"f706e096-2b8c-4696-a52f-a2e200508170\") " Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.338055 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f706e096-2b8c-4696-a52f-a2e200508170-kube-api-access-r4x8g" (OuterVolumeSpecName: "kube-api-access-r4x8g") pod "f706e096-2b8c-4696-a52f-a2e200508170" (UID: "f706e096-2b8c-4696-a52f-a2e200508170"). InnerVolumeSpecName "kube-api-access-r4x8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.338584 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f706e096-2b8c-4696-a52f-a2e200508170" (UID: "f706e096-2b8c-4696-a52f-a2e200508170"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.345964 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-scripts" (OuterVolumeSpecName: "scripts") pod "f706e096-2b8c-4696-a52f-a2e200508170" (UID: "f706e096-2b8c-4696-a52f-a2e200508170"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.361547 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f706e096-2b8c-4696-a52f-a2e200508170" (UID: "f706e096-2b8c-4696-a52f-a2e200508170"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.365108 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f706e096-2b8c-4696-a52f-a2e200508170" (UID: "f706e096-2b8c-4696-a52f-a2e200508170"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.368526 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-config-data" (OuterVolumeSpecName: "config-data") pod "f706e096-2b8c-4696-a52f-a2e200508170" (UID: "f706e096-2b8c-4696-a52f-a2e200508170"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.432540 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.432569 4646 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.432676 4646 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.432691 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.432704 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4x8g\" (UniqueName: \"kubernetes.io/projected/f706e096-2b8c-4696-a52f-a2e200508170-kube-api-access-r4x8g\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.432714 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f706e096-2b8c-4696-a52f-a2e200508170-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.619462 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jfdns" event={"ID":"f706e096-2b8c-4696-a52f-a2e200508170","Type":"ContainerDied","Data":"3b4906c367df01a988d432c924c22ff85f547f80da27e542df7d8c374da140ff"} Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.619505 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b4906c367df01a988d432c924c22ff85f547f80da27e542df7d8c374da140ff" Dec 03 11:12:49 crc kubenswrapper[4646]: I1203 11:12:49.619572 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jfdns" Dec 03 11:12:49 crc kubenswrapper[4646]: E1203 11:12:49.777905 4646 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf706e096_2b8c_4696_a52f_a2e200508170.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf706e096_2b8c_4696_a52f_a2e200508170.slice/crio-3b4906c367df01a988d432c924c22ff85f547f80da27e542df7d8c374da140ff\": RecentStats: unable to find data in memory cache]" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.473723 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-jfdns"] Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.481936 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-jfdns"] Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.562882 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-92v4c"] Dec 03 11:12:50 crc kubenswrapper[4646]: E1203 11:12:50.563542 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f706e096-2b8c-4696-a52f-a2e200508170" containerName="keystone-bootstrap" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.569633 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f706e096-2b8c-4696-a52f-a2e200508170" containerName="keystone-bootstrap" Dec 03 11:12:50 crc kubenswrapper[4646]: E1203 11:12:50.569698 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913b988e-bed2-4d69-90af-72d397584c9f" containerName="init" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.569709 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="913b988e-bed2-4d69-90af-72d397584c9f" containerName="init" Dec 03 11:12:50 crc kubenswrapper[4646]: E1203 11:12:50.569715 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8e61d88-5b29-4886-80aa-3a818c0594c6" containerName="init" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.569723 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8e61d88-5b29-4886-80aa-3a818c0594c6" containerName="init" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.570046 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="913b988e-bed2-4d69-90af-72d397584c9f" containerName="init" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.570066 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8e61d88-5b29-4886-80aa-3a818c0594c6" containerName="init" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.570080 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="f706e096-2b8c-4696-a52f-a2e200508170" containerName="keystone-bootstrap" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.570647 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.574746 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ddjxq" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.574795 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.574932 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.574942 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.575157 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.587072 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-config-data\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.587224 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-scripts\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.587383 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-credential-keys\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.587492 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94fb7\" (UniqueName: \"kubernetes.io/projected/36606312-32d1-4c3a-be99-5477047af153-kube-api-access-94fb7\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.587589 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-combined-ca-bundle\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.587670 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-fernet-keys\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.593075 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-92v4c"] Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.688734 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-combined-ca-bundle\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.688791 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-fernet-keys\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.688867 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-config-data\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.688891 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-scripts\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.688950 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-credential-keys\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.688991 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94fb7\" (UniqueName: \"kubernetes.io/projected/36606312-32d1-4c3a-be99-5477047af153-kube-api-access-94fb7\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.694394 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-credential-keys\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.695301 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-scripts\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.696135 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-fernet-keys\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.698207 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-combined-ca-bundle\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.705802 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-config-data\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.710935 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94fb7\" (UniqueName: \"kubernetes.io/projected/36606312-32d1-4c3a-be99-5477047af153-kube-api-access-94fb7\") pod \"keystone-bootstrap-92v4c\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:50 crc kubenswrapper[4646]: I1203 11:12:50.889738 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:12:51 crc kubenswrapper[4646]: I1203 11:12:51.862883 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f706e096-2b8c-4696-a52f-a2e200508170" path="/var/lib/kubelet/pods/f706e096-2b8c-4696-a52f-a2e200508170/volumes" Dec 03 11:12:52 crc kubenswrapper[4646]: E1203 11:12:52.241228 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 03 11:12:52 crc kubenswrapper[4646]: E1203 11:12:52.241430 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f6rd9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-h4hq8_openstack(411d5bc2-b69a-4b93-bb7e-c8990b44e79a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:12:52 crc kubenswrapper[4646]: E1203 11:12:52.242609 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-h4hq8" podUID="411d5bc2-b69a-4b93-bb7e-c8990b44e79a" Dec 03 11:12:52 crc kubenswrapper[4646]: E1203 11:12:52.562116 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 03 11:12:52 crc kubenswrapper[4646]: E1203 11:12:52.562640 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5fchb5h5cch679h654h574h5fdh57fh675h699h587h5d5h599h7bh644h6h5fh567hcfh646hch5b7h5f9h66dh689h66ch5fch694h9h87hc5h577q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2cx9n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(7715c927-8035-48d3-908e-4f3838f296aa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:12:52 crc kubenswrapper[4646]: E1203 11:12:52.649153 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-h4hq8" podUID="411d5bc2-b69a-4b93-bb7e-c8990b44e79a" Dec 03 11:12:57 crc kubenswrapper[4646]: I1203 11:12:57.658917 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-tqqfs" podUID="758d0aa0-8428-46bb-902d-f751b670c1ac" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: i/o timeout" Dec 03 11:12:59 crc kubenswrapper[4646]: I1203 11:12:59.915455 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.051654 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-ovsdbserver-nb\") pod \"758d0aa0-8428-46bb-902d-f751b670c1ac\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.051697 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-ovsdbserver-sb\") pod \"758d0aa0-8428-46bb-902d-f751b670c1ac\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.051735 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-config\") pod \"758d0aa0-8428-46bb-902d-f751b670c1ac\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.051782 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-dns-svc\") pod \"758d0aa0-8428-46bb-902d-f751b670c1ac\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.051814 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dhxv\" (UniqueName: \"kubernetes.io/projected/758d0aa0-8428-46bb-902d-f751b670c1ac-kube-api-access-7dhxv\") pod \"758d0aa0-8428-46bb-902d-f751b670c1ac\" (UID: \"758d0aa0-8428-46bb-902d-f751b670c1ac\") " Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.096486 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/758d0aa0-8428-46bb-902d-f751b670c1ac-kube-api-access-7dhxv" (OuterVolumeSpecName: "kube-api-access-7dhxv") pod "758d0aa0-8428-46bb-902d-f751b670c1ac" (UID: "758d0aa0-8428-46bb-902d-f751b670c1ac"). InnerVolumeSpecName "kube-api-access-7dhxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.124631 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "758d0aa0-8428-46bb-902d-f751b670c1ac" (UID: "758d0aa0-8428-46bb-902d-f751b670c1ac"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.130010 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "758d0aa0-8428-46bb-902d-f751b670c1ac" (UID: "758d0aa0-8428-46bb-902d-f751b670c1ac"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.131005 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "758d0aa0-8428-46bb-902d-f751b670c1ac" (UID: "758d0aa0-8428-46bb-902d-f751b670c1ac"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.160311 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dhxv\" (UniqueName: \"kubernetes.io/projected/758d0aa0-8428-46bb-902d-f751b670c1ac-kube-api-access-7dhxv\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.160355 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.160363 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.160373 4646 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.205531 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-config" (OuterVolumeSpecName: "config") pod "758d0aa0-8428-46bb-902d-f751b670c1ac" (UID: "758d0aa0-8428-46bb-902d-f751b670c1ac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.261861 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/758d0aa0-8428-46bb-902d-f751b670c1ac-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.750687 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-tqqfs" event={"ID":"758d0aa0-8428-46bb-902d-f751b670c1ac","Type":"ContainerDied","Data":"8368c35b17ca70b9803582f0023161a074fe71c1026bbdc5de27e10585b92ec0"} Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.750749 4646 scope.go:117] "RemoveContainer" containerID="8adc4d7cca6da85aa4a5b3e75931c2edb7f7d3204c30b6e47348f912816cb87f" Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.750792 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-tqqfs" Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.789135 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-tqqfs"] Dec 03 11:13:00 crc kubenswrapper[4646]: I1203 11:13:00.797188 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-tqqfs"] Dec 03 11:13:01 crc kubenswrapper[4646]: I1203 11:13:01.040973 4646 scope.go:117] "RemoveContainer" containerID="3a09cbb4ef5c13779b35dbbb2e555bfad93b0a9b0ccb80f40cd03bc9dc1c1f2d" Dec 03 11:13:01 crc kubenswrapper[4646]: E1203 11:13:01.069197 4646 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 03 11:13:01 crc kubenswrapper[4646]: E1203 11:13:01.069372 4646 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vzqhk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-rtwcp_openstack(76443d6c-7930-4c62-abbb-d7ff831232ca): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 11:13:01 crc kubenswrapper[4646]: E1203 11:13:01.070627 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-rtwcp" podUID="76443d6c-7930-4c62-abbb-d7ff831232ca" Dec 03 11:13:01 crc kubenswrapper[4646]: I1203 11:13:01.458669 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-92v4c"] Dec 03 11:13:01 crc kubenswrapper[4646]: W1203 11:13:01.475879 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36606312_32d1_4c3a_be99_5477047af153.slice/crio-ec26d46a78f030f87046b09a651119b22e4babf8c609c7451ddb4ee13f576560 WatchSource:0}: Error finding container ec26d46a78f030f87046b09a651119b22e4babf8c609c7451ddb4ee13f576560: Status 404 returned error can't find the container with id ec26d46a78f030f87046b09a651119b22e4babf8c609c7451ddb4ee13f576560 Dec 03 11:13:01 crc kubenswrapper[4646]: I1203 11:13:01.761078 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7715c927-8035-48d3-908e-4f3838f296aa","Type":"ContainerStarted","Data":"487ee5827c75a4af2682b7272eaee399afa12a6e73d723eae253df638228a6af"} Dec 03 11:13:01 crc kubenswrapper[4646]: I1203 11:13:01.762975 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ghphj" event={"ID":"c0028dab-487a-4f33-be6b-0f082d4f617e","Type":"ContainerStarted","Data":"a479fcfefd601ad29d040fd8ef9c169043f069b11742878431c036cead9d06d6"} Dec 03 11:13:01 crc kubenswrapper[4646]: I1203 11:13:01.765402 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-92v4c" event={"ID":"36606312-32d1-4c3a-be99-5477047af153","Type":"ContainerStarted","Data":"d541474f3e6ba25e699284ba5c49a45e9a46fb9220c76ed37b6440435a73e182"} Dec 03 11:13:01 crc kubenswrapper[4646]: I1203 11:13:01.765434 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-92v4c" event={"ID":"36606312-32d1-4c3a-be99-5477047af153","Type":"ContainerStarted","Data":"ec26d46a78f030f87046b09a651119b22e4babf8c609c7451ddb4ee13f576560"} Dec 03 11:13:01 crc kubenswrapper[4646]: E1203 11:13:01.766687 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-rtwcp" podUID="76443d6c-7930-4c62-abbb-d7ff831232ca" Dec 03 11:13:01 crc kubenswrapper[4646]: I1203 11:13:01.816276 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-ghphj" podStartSLOduration=2.904592543 podStartE2EDuration="28.816257997s" podCreationTimestamp="2025-12-03 11:12:33 +0000 UTC" firstStartedPulling="2025-12-03 11:12:35.117231556 +0000 UTC m=+1131.580287691" lastFinishedPulling="2025-12-03 11:13:01.02889701 +0000 UTC m=+1157.491953145" observedRunningTime="2025-12-03 11:13:01.791545666 +0000 UTC m=+1158.254601791" watchObservedRunningTime="2025-12-03 11:13:01.816257997 +0000 UTC m=+1158.279314132" Dec 03 11:13:01 crc kubenswrapper[4646]: I1203 11:13:01.818876 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-92v4c" podStartSLOduration=11.818866937 podStartE2EDuration="11.818866937s" podCreationTimestamp="2025-12-03 11:12:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:13:01.815860326 +0000 UTC m=+1158.278916481" watchObservedRunningTime="2025-12-03 11:13:01.818866937 +0000 UTC m=+1158.281923062" Dec 03 11:13:01 crc kubenswrapper[4646]: I1203 11:13:01.858945 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="758d0aa0-8428-46bb-902d-f751b670c1ac" path="/var/lib/kubelet/pods/758d0aa0-8428-46bb-902d-f751b670c1ac/volumes" Dec 03 11:13:02 crc kubenswrapper[4646]: I1203 11:13:02.660404 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-tqqfs" podUID="758d0aa0-8428-46bb-902d-f751b670c1ac" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: i/o timeout" Dec 03 11:13:02 crc kubenswrapper[4646]: I1203 11:13:02.778753 4646 generic.go:334] "Generic (PLEG): container finished" podID="aad954e4-5d92-4b2b-9771-a5f78f11d169" containerID="0b73033a13582c0cbbc6b90aa28c4838f5894f967db61e4e58282e22bebaab17" exitCode=0 Dec 03 11:13:02 crc kubenswrapper[4646]: I1203 11:13:02.778825 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2nh9f" event={"ID":"aad954e4-5d92-4b2b-9771-a5f78f11d169","Type":"ContainerDied","Data":"0b73033a13582c0cbbc6b90aa28c4838f5894f967db61e4e58282e22bebaab17"} Dec 03 11:13:03 crc kubenswrapper[4646]: I1203 11:13:03.788899 4646 generic.go:334] "Generic (PLEG): container finished" podID="c0028dab-487a-4f33-be6b-0f082d4f617e" containerID="a479fcfefd601ad29d040fd8ef9c169043f069b11742878431c036cead9d06d6" exitCode=0 Dec 03 11:13:03 crc kubenswrapper[4646]: I1203 11:13:03.788971 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ghphj" event={"ID":"c0028dab-487a-4f33-be6b-0f082d4f617e","Type":"ContainerDied","Data":"a479fcfefd601ad29d040fd8ef9c169043f069b11742878431c036cead9d06d6"} Dec 03 11:13:04 crc kubenswrapper[4646]: I1203 11:13:04.798472 4646 generic.go:334] "Generic (PLEG): container finished" podID="36606312-32d1-4c3a-be99-5477047af153" containerID="d541474f3e6ba25e699284ba5c49a45e9a46fb9220c76ed37b6440435a73e182" exitCode=0 Dec 03 11:13:04 crc kubenswrapper[4646]: I1203 11:13:04.798663 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-92v4c" event={"ID":"36606312-32d1-4c3a-be99-5477047af153","Type":"ContainerDied","Data":"d541474f3e6ba25e699284ba5c49a45e9a46fb9220c76ed37b6440435a73e182"} Dec 03 11:13:06 crc kubenswrapper[4646]: I1203 11:13:06.828297 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ghphj" event={"ID":"c0028dab-487a-4f33-be6b-0f082d4f617e","Type":"ContainerDied","Data":"c55e11fa607664f234091d33313c8cf6bd4c6f8a4d93458e7736775908083555"} Dec 03 11:13:06 crc kubenswrapper[4646]: I1203 11:13:06.828934 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c55e11fa607664f234091d33313c8cf6bd4c6f8a4d93458e7736775908083555" Dec 03 11:13:06 crc kubenswrapper[4646]: I1203 11:13:06.830968 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2nh9f" event={"ID":"aad954e4-5d92-4b2b-9771-a5f78f11d169","Type":"ContainerDied","Data":"e1b8411e617482fad0b1b766802894be967bdb8788000ca6e634171ff64f7bb1"} Dec 03 11:13:06 crc kubenswrapper[4646]: I1203 11:13:06.831061 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1b8411e617482fad0b1b766802894be967bdb8788000ca6e634171ff64f7bb1" Dec 03 11:13:06 crc kubenswrapper[4646]: I1203 11:13:06.833013 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-92v4c" event={"ID":"36606312-32d1-4c3a-be99-5477047af153","Type":"ContainerDied","Data":"ec26d46a78f030f87046b09a651119b22e4babf8c609c7451ddb4ee13f576560"} Dec 03 11:13:06 crc kubenswrapper[4646]: I1203 11:13:06.833038 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec26d46a78f030f87046b09a651119b22e4babf8c609c7451ddb4ee13f576560" Dec 03 11:13:06 crc kubenswrapper[4646]: I1203 11:13:06.861088 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ghphj" Dec 03 11:13:06 crc kubenswrapper[4646]: I1203 11:13:06.931520 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:13:06 crc kubenswrapper[4646]: I1203 11:13:06.946912 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2nh9f" Dec 03 11:13:06 crc kubenswrapper[4646]: I1203 11:13:06.997106 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-scripts\") pod \"c0028dab-487a-4f33-be6b-0f082d4f617e\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " Dec 03 11:13:06 crc kubenswrapper[4646]: I1203 11:13:06.997213 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-combined-ca-bundle\") pod \"c0028dab-487a-4f33-be6b-0f082d4f617e\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " Dec 03 11:13:06 crc kubenswrapper[4646]: I1203 11:13:06.997305 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trtqv\" (UniqueName: \"kubernetes.io/projected/c0028dab-487a-4f33-be6b-0f082d4f617e-kube-api-access-trtqv\") pod \"c0028dab-487a-4f33-be6b-0f082d4f617e\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " Dec 03 11:13:06 crc kubenswrapper[4646]: I1203 11:13:06.997355 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-config-data\") pod \"c0028dab-487a-4f33-be6b-0f082d4f617e\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " Dec 03 11:13:06 crc kubenswrapper[4646]: I1203 11:13:06.997409 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0028dab-487a-4f33-be6b-0f082d4f617e-logs\") pod \"c0028dab-487a-4f33-be6b-0f082d4f617e\" (UID: \"c0028dab-487a-4f33-be6b-0f082d4f617e\") " Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.001021 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0028dab-487a-4f33-be6b-0f082d4f617e-logs" (OuterVolumeSpecName: "logs") pod "c0028dab-487a-4f33-be6b-0f082d4f617e" (UID: "c0028dab-487a-4f33-be6b-0f082d4f617e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.011516 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-scripts" (OuterVolumeSpecName: "scripts") pod "c0028dab-487a-4f33-be6b-0f082d4f617e" (UID: "c0028dab-487a-4f33-be6b-0f082d4f617e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.019031 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0028dab-487a-4f33-be6b-0f082d4f617e-kube-api-access-trtqv" (OuterVolumeSpecName: "kube-api-access-trtqv") pod "c0028dab-487a-4f33-be6b-0f082d4f617e" (UID: "c0028dab-487a-4f33-be6b-0f082d4f617e"). InnerVolumeSpecName "kube-api-access-trtqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.036210 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-config-data" (OuterVolumeSpecName: "config-data") pod "c0028dab-487a-4f33-be6b-0f082d4f617e" (UID: "c0028dab-487a-4f33-be6b-0f082d4f617e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.049084 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0028dab-487a-4f33-be6b-0f082d4f617e" (UID: "c0028dab-487a-4f33-be6b-0f082d4f617e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.099219 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94fb7\" (UniqueName: \"kubernetes.io/projected/36606312-32d1-4c3a-be99-5477047af153-kube-api-access-94fb7\") pod \"36606312-32d1-4c3a-be99-5477047af153\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.099279 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aad954e4-5d92-4b2b-9771-a5f78f11d169-config\") pod \"aad954e4-5d92-4b2b-9771-a5f78f11d169\" (UID: \"aad954e4-5d92-4b2b-9771-a5f78f11d169\") " Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.099312 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-fernet-keys\") pod \"36606312-32d1-4c3a-be99-5477047af153\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.099350 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-combined-ca-bundle\") pod \"36606312-32d1-4c3a-be99-5477047af153\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.099377 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad954e4-5d92-4b2b-9771-a5f78f11d169-combined-ca-bundle\") pod \"aad954e4-5d92-4b2b-9771-a5f78f11d169\" (UID: \"aad954e4-5d92-4b2b-9771-a5f78f11d169\") " Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.099428 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh5xk\" (UniqueName: \"kubernetes.io/projected/aad954e4-5d92-4b2b-9771-a5f78f11d169-kube-api-access-lh5xk\") pod \"aad954e4-5d92-4b2b-9771-a5f78f11d169\" (UID: \"aad954e4-5d92-4b2b-9771-a5f78f11d169\") " Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.099507 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-config-data\") pod \"36606312-32d1-4c3a-be99-5477047af153\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.099525 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-scripts\") pod \"36606312-32d1-4c3a-be99-5477047af153\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.099569 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-credential-keys\") pod \"36606312-32d1-4c3a-be99-5477047af153\" (UID: \"36606312-32d1-4c3a-be99-5477047af153\") " Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.099887 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.099898 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.099909 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trtqv\" (UniqueName: \"kubernetes.io/projected/c0028dab-487a-4f33-be6b-0f082d4f617e-kube-api-access-trtqv\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.099917 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0028dab-487a-4f33-be6b-0f082d4f617e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.099925 4646 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0028dab-487a-4f33-be6b-0f082d4f617e-logs\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.102767 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36606312-32d1-4c3a-be99-5477047af153-kube-api-access-94fb7" (OuterVolumeSpecName: "kube-api-access-94fb7") pod "36606312-32d1-4c3a-be99-5477047af153" (UID: "36606312-32d1-4c3a-be99-5477047af153"). InnerVolumeSpecName "kube-api-access-94fb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.114086 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aad954e4-5d92-4b2b-9771-a5f78f11d169-kube-api-access-lh5xk" (OuterVolumeSpecName: "kube-api-access-lh5xk") pod "aad954e4-5d92-4b2b-9771-a5f78f11d169" (UID: "aad954e4-5d92-4b2b-9771-a5f78f11d169"). InnerVolumeSpecName "kube-api-access-lh5xk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.114092 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "36606312-32d1-4c3a-be99-5477047af153" (UID: "36606312-32d1-4c3a-be99-5477047af153"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.114307 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "36606312-32d1-4c3a-be99-5477047af153" (UID: "36606312-32d1-4c3a-be99-5477047af153"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.117563 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-scripts" (OuterVolumeSpecName: "scripts") pod "36606312-32d1-4c3a-be99-5477047af153" (UID: "36606312-32d1-4c3a-be99-5477047af153"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.127420 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad954e4-5d92-4b2b-9771-a5f78f11d169-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aad954e4-5d92-4b2b-9771-a5f78f11d169" (UID: "aad954e4-5d92-4b2b-9771-a5f78f11d169"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.128736 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad954e4-5d92-4b2b-9771-a5f78f11d169-config" (OuterVolumeSpecName: "config") pod "aad954e4-5d92-4b2b-9771-a5f78f11d169" (UID: "aad954e4-5d92-4b2b-9771-a5f78f11d169"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.131462 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-config-data" (OuterVolumeSpecName: "config-data") pod "36606312-32d1-4c3a-be99-5477047af153" (UID: "36606312-32d1-4c3a-be99-5477047af153"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.131518 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36606312-32d1-4c3a-be99-5477047af153" (UID: "36606312-32d1-4c3a-be99-5477047af153"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.201160 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.201188 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.201197 4646 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.201210 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94fb7\" (UniqueName: \"kubernetes.io/projected/36606312-32d1-4c3a-be99-5477047af153-kube-api-access-94fb7\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.201220 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/aad954e4-5d92-4b2b-9771-a5f78f11d169-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.201228 4646 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.201236 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36606312-32d1-4c3a-be99-5477047af153-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.201244 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad954e4-5d92-4b2b-9771-a5f78f11d169-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.201252 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh5xk\" (UniqueName: \"kubernetes.io/projected/aad954e4-5d92-4b2b-9771-a5f78f11d169-kube-api-access-lh5xk\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.844527 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h4hq8" event={"ID":"411d5bc2-b69a-4b93-bb7e-c8990b44e79a","Type":"ContainerStarted","Data":"2b9c5c1bdd25667984f4878e0cb31ba5d3c44d04d285383a2d53bea7cb8c68a5"} Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.849946 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ghphj" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.852467 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-92v4c" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.853053 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2nh9f" Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.858230 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7715c927-8035-48d3-908e-4f3838f296aa","Type":"ContainerStarted","Data":"d9b3e1adfd5004498124ec7e0fc51643711a2c085b29b82477eac6993d0c4ee6"} Dec 03 11:13:07 crc kubenswrapper[4646]: I1203 11:13:07.875951 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-h4hq8" podStartSLOduration=3.359899403 podStartE2EDuration="34.875922525s" podCreationTimestamp="2025-12-03 11:12:33 +0000 UTC" firstStartedPulling="2025-12-03 11:12:35.130681776 +0000 UTC m=+1131.593737911" lastFinishedPulling="2025-12-03 11:13:06.646704908 +0000 UTC m=+1163.109761033" observedRunningTime="2025-12-03 11:13:07.875810922 +0000 UTC m=+1164.338867057" watchObservedRunningTime="2025-12-03 11:13:07.875922525 +0000 UTC m=+1164.338978680" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.043416 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-96574894d-gjvg2"] Dec 03 11:13:08 crc kubenswrapper[4646]: E1203 11:13:08.043745 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="758d0aa0-8428-46bb-902d-f751b670c1ac" containerName="dnsmasq-dns" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.043759 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="758d0aa0-8428-46bb-902d-f751b670c1ac" containerName="dnsmasq-dns" Dec 03 11:13:08 crc kubenswrapper[4646]: E1203 11:13:08.043769 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="758d0aa0-8428-46bb-902d-f751b670c1ac" containerName="init" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.043775 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="758d0aa0-8428-46bb-902d-f751b670c1ac" containerName="init" Dec 03 11:13:08 crc kubenswrapper[4646]: E1203 11:13:08.043789 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0028dab-487a-4f33-be6b-0f082d4f617e" containerName="placement-db-sync" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.043795 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0028dab-487a-4f33-be6b-0f082d4f617e" containerName="placement-db-sync" Dec 03 11:13:08 crc kubenswrapper[4646]: E1203 11:13:08.043809 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aad954e4-5d92-4b2b-9771-a5f78f11d169" containerName="neutron-db-sync" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.043814 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="aad954e4-5d92-4b2b-9771-a5f78f11d169" containerName="neutron-db-sync" Dec 03 11:13:08 crc kubenswrapper[4646]: E1203 11:13:08.043822 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36606312-32d1-4c3a-be99-5477047af153" containerName="keystone-bootstrap" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.043827 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="36606312-32d1-4c3a-be99-5477047af153" containerName="keystone-bootstrap" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.043974 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="aad954e4-5d92-4b2b-9771-a5f78f11d169" containerName="neutron-db-sync" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.043984 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="758d0aa0-8428-46bb-902d-f751b670c1ac" containerName="dnsmasq-dns" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.044004 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="36606312-32d1-4c3a-be99-5477047af153" containerName="keystone-bootstrap" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.044012 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0028dab-487a-4f33-be6b-0f082d4f617e" containerName="placement-db-sync" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.044844 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.050461 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.051438 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.051754 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.051903 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-g4pdh" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.052777 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.063714 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-96574894d-gjvg2"] Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.212152 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-779dc4bcb9-gdhmd"] Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.215724 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.218296 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b862ee56-bca1-4ea4-b2e8-d1ef47365550-combined-ca-bundle\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.218370 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b862ee56-bca1-4ea4-b2e8-d1ef47365550-internal-tls-certs\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.218399 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b862ee56-bca1-4ea4-b2e8-d1ef47365550-scripts\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.218489 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b862ee56-bca1-4ea4-b2e8-d1ef47365550-config-data\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.218514 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b862ee56-bca1-4ea4-b2e8-d1ef47365550-public-tls-certs\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.218600 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b862ee56-bca1-4ea4-b2e8-d1ef47365550-logs\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.218627 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5b58\" (UniqueName: \"kubernetes.io/projected/b862ee56-bca1-4ea4-b2e8-d1ef47365550-kube-api-access-h5b58\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.251424 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.251574 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ddjxq" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.251424 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.251918 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.252727 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.254743 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.307455 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-779dc4bcb9-gdhmd"] Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.324233 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-credential-keys\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.324298 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b862ee56-bca1-4ea4-b2e8-d1ef47365550-combined-ca-bundle\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.324348 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b862ee56-bca1-4ea4-b2e8-d1ef47365550-internal-tls-certs\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.324376 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b862ee56-bca1-4ea4-b2e8-d1ef47365550-scripts\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.324401 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-config-data\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.324463 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-fernet-keys\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.324487 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk9v9\" (UniqueName: \"kubernetes.io/projected/7f051542-4055-4cde-aaf3-8c30075cb832-kube-api-access-jk9v9\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.324507 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-public-tls-certs\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.324536 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b862ee56-bca1-4ea4-b2e8-d1ef47365550-config-data\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.324555 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-internal-tls-certs\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.324575 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b862ee56-bca1-4ea4-b2e8-d1ef47365550-public-tls-certs\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.324633 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-scripts\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.324660 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-combined-ca-bundle\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.324688 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b862ee56-bca1-4ea4-b2e8-d1ef47365550-logs\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.324712 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5b58\" (UniqueName: \"kubernetes.io/projected/b862ee56-bca1-4ea4-b2e8-d1ef47365550-kube-api-access-h5b58\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.332699 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b862ee56-bca1-4ea4-b2e8-d1ef47365550-internal-tls-certs\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.355802 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b862ee56-bca1-4ea4-b2e8-d1ef47365550-logs\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.373092 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b862ee56-bca1-4ea4-b2e8-d1ef47365550-combined-ca-bundle\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.373554 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b862ee56-bca1-4ea4-b2e8-d1ef47365550-public-tls-certs\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.386458 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b862ee56-bca1-4ea4-b2e8-d1ef47365550-config-data\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.394724 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b862ee56-bca1-4ea4-b2e8-d1ef47365550-scripts\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.426047 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-7j5cd"] Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.426137 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-scripts\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.426200 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-combined-ca-bundle\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.426257 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-credential-keys\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.426286 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-config-data\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.426350 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-fernet-keys\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.426370 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk9v9\" (UniqueName: \"kubernetes.io/projected/7f051542-4055-4cde-aaf3-8c30075cb832-kube-api-access-jk9v9\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.426401 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-public-tls-certs\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.426432 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-internal-tls-certs\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.427709 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.433022 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-7j5cd"] Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.440762 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5b58\" (UniqueName: \"kubernetes.io/projected/b862ee56-bca1-4ea4-b2e8-d1ef47365550-kube-api-access-h5b58\") pod \"placement-96574894d-gjvg2\" (UID: \"b862ee56-bca1-4ea4-b2e8-d1ef47365550\") " pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.466006 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-internal-tls-certs\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.473710 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-scripts\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.474293 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-public-tls-certs\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.480603 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk9v9\" (UniqueName: \"kubernetes.io/projected/7f051542-4055-4cde-aaf3-8c30075cb832-kube-api-access-jk9v9\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.492132 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-combined-ca-bundle\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.493190 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-config-data\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.530904 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-config\") pod \"dnsmasq-dns-5f66db59b9-7j5cd\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.530963 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-ovsdbserver-nb\") pod \"dnsmasq-dns-5f66db59b9-7j5cd\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.530996 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz525\" (UniqueName: \"kubernetes.io/projected/9aca34cf-ca0a-4449-9188-04d05766a266-kube-api-access-sz525\") pod \"dnsmasq-dns-5f66db59b9-7j5cd\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.531020 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-ovsdbserver-sb\") pod \"dnsmasq-dns-5f66db59b9-7j5cd\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.531067 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-dns-svc\") pod \"dnsmasq-dns-5f66db59b9-7j5cd\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.539432 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-credential-keys\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.571110 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7f051542-4055-4cde-aaf3-8c30075cb832-fernet-keys\") pod \"keystone-779dc4bcb9-gdhmd\" (UID: \"7f051542-4055-4cde-aaf3-8c30075cb832\") " pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.632045 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-ovsdbserver-sb\") pod \"dnsmasq-dns-5f66db59b9-7j5cd\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.632391 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-dns-svc\") pod \"dnsmasq-dns-5f66db59b9-7j5cd\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.632544 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-config\") pod \"dnsmasq-dns-5f66db59b9-7j5cd\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.632939 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-ovsdbserver-nb\") pod \"dnsmasq-dns-5f66db59b9-7j5cd\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.633055 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz525\" (UniqueName: \"kubernetes.io/projected/9aca34cf-ca0a-4449-9188-04d05766a266-kube-api-access-sz525\") pod \"dnsmasq-dns-5f66db59b9-7j5cd\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.634185 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-ovsdbserver-sb\") pod \"dnsmasq-dns-5f66db59b9-7j5cd\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.634866 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-dns-svc\") pod \"dnsmasq-dns-5f66db59b9-7j5cd\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.635879 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-config\") pod \"dnsmasq-dns-5f66db59b9-7j5cd\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.636507 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-ovsdbserver-nb\") pod \"dnsmasq-dns-5f66db59b9-7j5cd\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.664277 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-9f5b87b8-m9k27"] Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.665810 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.714071 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz525\" (UniqueName: \"kubernetes.io/projected/9aca34cf-ca0a-4449-9188-04d05766a266-kube-api-access-sz525\") pod \"dnsmasq-dns-5f66db59b9-7j5cd\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.723805 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-2zxxm" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.724252 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.724554 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.724849 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.736943 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-ovndb-tls-certs\") pod \"neutron-9f5b87b8-m9k27\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.737713 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-httpd-config\") pod \"neutron-9f5b87b8-m9k27\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.737826 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5jk6\" (UniqueName: \"kubernetes.io/projected/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-kube-api-access-v5jk6\") pod \"neutron-9f5b87b8-m9k27\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.737923 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-combined-ca-bundle\") pod \"neutron-9f5b87b8-m9k27\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.738007 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-config\") pod \"neutron-9f5b87b8-m9k27\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.737586 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.748914 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9f5b87b8-m9k27"] Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.845898 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.846093 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-combined-ca-bundle\") pod \"neutron-9f5b87b8-m9k27\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.847422 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-config\") pod \"neutron-9f5b87b8-m9k27\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.847726 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-ovndb-tls-certs\") pod \"neutron-9f5b87b8-m9k27\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.847819 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-httpd-config\") pod \"neutron-9f5b87b8-m9k27\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.847849 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5jk6\" (UniqueName: \"kubernetes.io/projected/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-kube-api-access-v5jk6\") pod \"neutron-9f5b87b8-m9k27\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.850191 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-combined-ca-bundle\") pod \"neutron-9f5b87b8-m9k27\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.853105 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-ovndb-tls-certs\") pod \"neutron-9f5b87b8-m9k27\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.854387 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-config\") pod \"neutron-9f5b87b8-m9k27\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.895923 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-httpd-config\") pod \"neutron-9f5b87b8-m9k27\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.896773 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5jk6\" (UniqueName: \"kubernetes.io/projected/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-kube-api-access-v5jk6\") pod \"neutron-9f5b87b8-m9k27\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:08 crc kubenswrapper[4646]: I1203 11:13:08.925219 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:09 crc kubenswrapper[4646]: I1203 11:13:09.058784 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:09 crc kubenswrapper[4646]: I1203 11:13:09.311366 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-779dc4bcb9-gdhmd"] Dec 03 11:13:09 crc kubenswrapper[4646]: I1203 11:13:09.623273 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-96574894d-gjvg2"] Dec 03 11:13:09 crc kubenswrapper[4646]: I1203 11:13:09.750447 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-7j5cd"] Dec 03 11:13:09 crc kubenswrapper[4646]: I1203 11:13:09.929362 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" event={"ID":"9aca34cf-ca0a-4449-9188-04d05766a266","Type":"ContainerStarted","Data":"8c6ad3f743df7923262c547b799ae6a01f3939d0fad68f7a0db8df911cfad517"} Dec 03 11:13:09 crc kubenswrapper[4646]: I1203 11:13:09.930618 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-96574894d-gjvg2" event={"ID":"b862ee56-bca1-4ea4-b2e8-d1ef47365550","Type":"ContainerStarted","Data":"913f9ae32587916f7cb22c9441cb49ccd139b6da859c66191f46e964cbe6e743"} Dec 03 11:13:09 crc kubenswrapper[4646]: I1203 11:13:09.933283 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-779dc4bcb9-gdhmd" event={"ID":"7f051542-4055-4cde-aaf3-8c30075cb832","Type":"ContainerStarted","Data":"af7ffb872ac7c6472d6bf23ab6df00149d891f7be617858062f6065516a969f8"} Dec 03 11:13:09 crc kubenswrapper[4646]: I1203 11:13:09.933309 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-779dc4bcb9-gdhmd" event={"ID":"7f051542-4055-4cde-aaf3-8c30075cb832","Type":"ContainerStarted","Data":"f324232b4c21f88588eb64e4ab09dc567492fee49a62dd87b63f6566a1ddd014"} Dec 03 11:13:09 crc kubenswrapper[4646]: I1203 11:13:09.936454 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:09 crc kubenswrapper[4646]: I1203 11:13:09.963031 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-779dc4bcb9-gdhmd" podStartSLOduration=1.963008284 podStartE2EDuration="1.963008284s" podCreationTimestamp="2025-12-03 11:13:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:13:09.960880947 +0000 UTC m=+1166.423937092" watchObservedRunningTime="2025-12-03 11:13:09.963008284 +0000 UTC m=+1166.426064419" Dec 03 11:13:09 crc kubenswrapper[4646]: I1203 11:13:09.990294 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9f5b87b8-m9k27"] Dec 03 11:13:09 crc kubenswrapper[4646]: W1203 11:13:09.997617 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0d65d94_d172_41e9_9e61_d44dd40dcb3e.slice/crio-0b8a20a4736f375e0bbe9e808ae10768ab044ed2193046e4e6183f86481c50af WatchSource:0}: Error finding container 0b8a20a4736f375e0bbe9e808ae10768ab044ed2193046e4e6183f86481c50af: Status 404 returned error can't find the container with id 0b8a20a4736f375e0bbe9e808ae10768ab044ed2193046e4e6183f86481c50af Dec 03 11:13:10 crc kubenswrapper[4646]: E1203 11:13:10.567907 4646 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9aca34cf_ca0a_4449_9188_04d05766a266.slice/crio-conmon-6645a7892ca5cacbea98ff06dcf01bce92f9e0bcbde52602865e8ba4ba824745.scope\": RecentStats: unable to find data in memory cache]" Dec 03 11:13:10 crc kubenswrapper[4646]: I1203 11:13:10.947137 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-96574894d-gjvg2" event={"ID":"b862ee56-bca1-4ea4-b2e8-d1ef47365550","Type":"ContainerStarted","Data":"22b95c49563fd4ca7359397d0f4a5675affc1e15d41c059920bcc707127ff4fe"} Dec 03 11:13:10 crc kubenswrapper[4646]: I1203 11:13:10.947479 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-96574894d-gjvg2" event={"ID":"b862ee56-bca1-4ea4-b2e8-d1ef47365550","Type":"ContainerStarted","Data":"e5563548d9c3f60dd1e8f2fd0f5b1b994cc671b47b593d0b4cd5894c8d462dc8"} Dec 03 11:13:10 crc kubenswrapper[4646]: I1203 11:13:10.947960 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:10 crc kubenswrapper[4646]: I1203 11:13:10.948013 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:10 crc kubenswrapper[4646]: I1203 11:13:10.956710 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9f5b87b8-m9k27" event={"ID":"d0d65d94-d172-41e9-9e61-d44dd40dcb3e","Type":"ContainerStarted","Data":"4b64af0103966c64399a94d2506ee2c8774c0cfe2acf8861a3d71d3ec4acd87e"} Dec 03 11:13:10 crc kubenswrapper[4646]: I1203 11:13:10.956765 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9f5b87b8-m9k27" event={"ID":"d0d65d94-d172-41e9-9e61-d44dd40dcb3e","Type":"ContainerStarted","Data":"0b8a20a4736f375e0bbe9e808ae10768ab044ed2193046e4e6183f86481c50af"} Dec 03 11:13:10 crc kubenswrapper[4646]: I1203 11:13:10.960466 4646 generic.go:334] "Generic (PLEG): container finished" podID="9aca34cf-ca0a-4449-9188-04d05766a266" containerID="6645a7892ca5cacbea98ff06dcf01bce92f9e0bcbde52602865e8ba4ba824745" exitCode=0 Dec 03 11:13:10 crc kubenswrapper[4646]: I1203 11:13:10.961838 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" event={"ID":"9aca34cf-ca0a-4449-9188-04d05766a266","Type":"ContainerDied","Data":"6645a7892ca5cacbea98ff06dcf01bce92f9e0bcbde52602865e8ba4ba824745"} Dec 03 11:13:10 crc kubenswrapper[4646]: I1203 11:13:10.985480 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-96574894d-gjvg2" podStartSLOduration=2.985458204 podStartE2EDuration="2.985458204s" podCreationTimestamp="2025-12-03 11:13:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:13:10.976406442 +0000 UTC m=+1167.439462577" watchObservedRunningTime="2025-12-03 11:13:10.985458204 +0000 UTC m=+1167.448514339" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.510266 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-59d75d495c-x9phz"] Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.512307 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.523031 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.523234 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.547097 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-59d75d495c-x9phz"] Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.599225 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rswgg\" (UniqueName: \"kubernetes.io/projected/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-kube-api-access-rswgg\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.599280 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-combined-ca-bundle\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.599311 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-public-tls-certs\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.599357 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-ovndb-tls-certs\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.599393 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-httpd-config\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.599423 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-config\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.599443 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-internal-tls-certs\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.700989 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-httpd-config\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.701048 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-config\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.701073 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-internal-tls-certs\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.701161 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rswgg\" (UniqueName: \"kubernetes.io/projected/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-kube-api-access-rswgg\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.701189 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-combined-ca-bundle\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.701214 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-public-tls-certs\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.701234 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-ovndb-tls-certs\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.712426 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-config\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.717454 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-combined-ca-bundle\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.718598 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-ovndb-tls-certs\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.719492 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-internal-tls-certs\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.720068 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-public-tls-certs\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.721135 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rswgg\" (UniqueName: \"kubernetes.io/projected/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-kube-api-access-rswgg\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.722987 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157-httpd-config\") pod \"neutron-59d75d495c-x9phz\" (UID: \"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157\") " pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:11 crc kubenswrapper[4646]: I1203 11:13:11.834820 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:12 crc kubenswrapper[4646]: I1203 11:13:12.397195 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-59d75d495c-x9phz"] Dec 03 11:13:13 crc kubenswrapper[4646]: I1203 11:13:13.005544 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59d75d495c-x9phz" event={"ID":"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157","Type":"ContainerStarted","Data":"fe53cebd2fc709de3f9dfe18214cf1c4d9d5f16f8771fcfd1f563b9f3e6ff910"} Dec 03 11:13:14 crc kubenswrapper[4646]: I1203 11:13:14.012825 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9f5b87b8-m9k27" event={"ID":"d0d65d94-d172-41e9-9e61-d44dd40dcb3e","Type":"ContainerStarted","Data":"4099cdb499116765b3041174066958a5a667f51efa1143d60d22231cb86a282e"} Dec 03 11:13:15 crc kubenswrapper[4646]: I1203 11:13:15.021768 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" event={"ID":"9aca34cf-ca0a-4449-9188-04d05766a266","Type":"ContainerStarted","Data":"9ecdf4d0d834d19420e14a6d02d8b43d5740ef4470fd5322a49d6828c6807b53"} Dec 03 11:13:15 crc kubenswrapper[4646]: I1203 11:13:15.022056 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:15 crc kubenswrapper[4646]: I1203 11:13:15.025199 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59d75d495c-x9phz" event={"ID":"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157","Type":"ContainerStarted","Data":"b7926579984c5a7d386ee3ee6d9e360de4698ba85a1bebfd1a80b996a128a78d"} Dec 03 11:13:15 crc kubenswrapper[4646]: I1203 11:13:15.025256 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59d75d495c-x9phz" event={"ID":"c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157","Type":"ContainerStarted","Data":"d2b8612eb97a04be3b11052baf646221e6c9fe848ccdfb7c1b3ce99d0b5d57bf"} Dec 03 11:13:15 crc kubenswrapper[4646]: I1203 11:13:15.025304 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:15 crc kubenswrapper[4646]: I1203 11:13:15.081561 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" podStartSLOduration=7.081538698 podStartE2EDuration="7.081538698s" podCreationTimestamp="2025-12-03 11:13:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:13:15.075610316 +0000 UTC m=+1171.538666451" watchObservedRunningTime="2025-12-03 11:13:15.081538698 +0000 UTC m=+1171.544594833" Dec 03 11:13:15 crc kubenswrapper[4646]: I1203 11:13:15.096107 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-9f5b87b8-m9k27" podStartSLOduration=7.096092015 podStartE2EDuration="7.096092015s" podCreationTimestamp="2025-12-03 11:13:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:13:15.094115611 +0000 UTC m=+1171.557171746" watchObservedRunningTime="2025-12-03 11:13:15.096092015 +0000 UTC m=+1171.559148150" Dec 03 11:13:15 crc kubenswrapper[4646]: I1203 11:13:15.226883 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-59d75d495c-x9phz" podStartSLOduration=4.226863177 podStartE2EDuration="4.226863177s" podCreationTimestamp="2025-12-03 11:13:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:13:15.220108052 +0000 UTC m=+1171.683164187" watchObservedRunningTime="2025-12-03 11:13:15.226863177 +0000 UTC m=+1171.689919312" Dec 03 11:13:17 crc kubenswrapper[4646]: I1203 11:13:17.056851 4646 generic.go:334] "Generic (PLEG): container finished" podID="411d5bc2-b69a-4b93-bb7e-c8990b44e79a" containerID="2b9c5c1bdd25667984f4878e0cb31ba5d3c44d04d285383a2d53bea7cb8c68a5" exitCode=0 Dec 03 11:13:17 crc kubenswrapper[4646]: I1203 11:13:17.057225 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h4hq8" event={"ID":"411d5bc2-b69a-4b93-bb7e-c8990b44e79a","Type":"ContainerDied","Data":"2b9c5c1bdd25667984f4878e0cb31ba5d3c44d04d285383a2d53bea7cb8c68a5"} Dec 03 11:13:19 crc kubenswrapper[4646]: I1203 11:13:19.488231 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h4hq8" Dec 03 11:13:19 crc kubenswrapper[4646]: I1203 11:13:19.682602 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-combined-ca-bundle\") pod \"411d5bc2-b69a-4b93-bb7e-c8990b44e79a\" (UID: \"411d5bc2-b69a-4b93-bb7e-c8990b44e79a\") " Dec 03 11:13:19 crc kubenswrapper[4646]: I1203 11:13:19.682666 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6rd9\" (UniqueName: \"kubernetes.io/projected/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-kube-api-access-f6rd9\") pod \"411d5bc2-b69a-4b93-bb7e-c8990b44e79a\" (UID: \"411d5bc2-b69a-4b93-bb7e-c8990b44e79a\") " Dec 03 11:13:19 crc kubenswrapper[4646]: I1203 11:13:19.682839 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-db-sync-config-data\") pod \"411d5bc2-b69a-4b93-bb7e-c8990b44e79a\" (UID: \"411d5bc2-b69a-4b93-bb7e-c8990b44e79a\") " Dec 03 11:13:19 crc kubenswrapper[4646]: I1203 11:13:19.687452 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "411d5bc2-b69a-4b93-bb7e-c8990b44e79a" (UID: "411d5bc2-b69a-4b93-bb7e-c8990b44e79a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:19 crc kubenswrapper[4646]: I1203 11:13:19.688491 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-kube-api-access-f6rd9" (OuterVolumeSpecName: "kube-api-access-f6rd9") pod "411d5bc2-b69a-4b93-bb7e-c8990b44e79a" (UID: "411d5bc2-b69a-4b93-bb7e-c8990b44e79a"). InnerVolumeSpecName "kube-api-access-f6rd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:13:19 crc kubenswrapper[4646]: I1203 11:13:19.710507 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "411d5bc2-b69a-4b93-bb7e-c8990b44e79a" (UID: "411d5bc2-b69a-4b93-bb7e-c8990b44e79a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:19 crc kubenswrapper[4646]: E1203 11:13:19.716812 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="7715c927-8035-48d3-908e-4f3838f296aa" Dec 03 11:13:19 crc kubenswrapper[4646]: I1203 11:13:19.785296 4646 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:19 crc kubenswrapper[4646]: I1203 11:13:19.785367 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:19 crc kubenswrapper[4646]: I1203 11:13:19.785380 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6rd9\" (UniqueName: \"kubernetes.io/projected/411d5bc2-b69a-4b93-bb7e-c8990b44e79a-kube-api-access-f6rd9\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.079583 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h4hq8" event={"ID":"411d5bc2-b69a-4b93-bb7e-c8990b44e79a","Type":"ContainerDied","Data":"2fb16059442c4c64e7e78a99fa95bfc4ad1e8e5472d3f0587adaa99dd2a83017"} Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.079631 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fb16059442c4c64e7e78a99fa95bfc4ad1e8e5472d3f0587adaa99dd2a83017" Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.079598 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h4hq8" Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.082584 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7715c927-8035-48d3-908e-4f3838f296aa","Type":"ContainerStarted","Data":"0f87fafd1a0ea0968db590d01faad64a686fc1d57ebaa88420f6e38a55f65ba2"} Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.082761 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7715c927-8035-48d3-908e-4f3838f296aa" containerName="ceilometer-notification-agent" containerID="cri-o://487ee5827c75a4af2682b7272eaee399afa12a6e73d723eae253df638228a6af" gracePeriod=30 Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.083030 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.083320 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7715c927-8035-48d3-908e-4f3838f296aa" containerName="proxy-httpd" containerID="cri-o://0f87fafd1a0ea0968db590d01faad64a686fc1d57ebaa88420f6e38a55f65ba2" gracePeriod=30 Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.083447 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7715c927-8035-48d3-908e-4f3838f296aa" containerName="sg-core" containerID="cri-o://d9b3e1adfd5004498124ec7e0fc51643711a2c085b29b82477eac6993d0c4ee6" gracePeriod=30 Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.087895 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rtwcp" event={"ID":"76443d6c-7930-4c62-abbb-d7ff831232ca","Type":"ContainerStarted","Data":"e7a283c6e0898451966968b2ef55a759b711684e8bdf653e930f30c52b97ed19"} Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.135778 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-rtwcp" podStartSLOduration=2.314563205 podStartE2EDuration="47.135763194s" podCreationTimestamp="2025-12-03 11:12:33 +0000 UTC" firstStartedPulling="2025-12-03 11:12:34.572486396 +0000 UTC m=+1131.035542531" lastFinishedPulling="2025-12-03 11:13:19.393686375 +0000 UTC m=+1175.856742520" observedRunningTime="2025-12-03 11:13:20.125765441 +0000 UTC m=+1176.588821596" watchObservedRunningTime="2025-12-03 11:13:20.135763194 +0000 UTC m=+1176.598819329" Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.933200 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-9f949dfbc-7rr98"] Dec 03 11:13:20 crc kubenswrapper[4646]: E1203 11:13:20.933902 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="411d5bc2-b69a-4b93-bb7e-c8990b44e79a" containerName="barbican-db-sync" Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.933921 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="411d5bc2-b69a-4b93-bb7e-c8990b44e79a" containerName="barbican-db-sync" Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.934102 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="411d5bc2-b69a-4b93-bb7e-c8990b44e79a" containerName="barbican-db-sync" Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.935010 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.939544 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.939737 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-x2m59" Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.939838 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.963532 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-9f949dfbc-7rr98"] Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.994806 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6b6bfd6548-5hldq"] Dec 03 11:13:20 crc kubenswrapper[4646]: I1203 11:13:20.996975 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.008193 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.009446 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6b6bfd6548-5hldq"] Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.019619 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-7j5cd"] Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.019907 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" podUID="9aca34cf-ca0a-4449-9188-04d05766a266" containerName="dnsmasq-dns" containerID="cri-o://9ecdf4d0d834d19420e14a6d02d8b43d5740ef4470fd5322a49d6828c6807b53" gracePeriod=10 Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.021526 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.046119 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaba04e4-bc4a-41a0-aa79-631669451754-combined-ca-bundle\") pod \"barbican-worker-9f949dfbc-7rr98\" (UID: \"eaba04e4-bc4a-41a0-aa79-631669451754\") " pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.046419 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaba04e4-bc4a-41a0-aa79-631669451754-config-data\") pod \"barbican-worker-9f949dfbc-7rr98\" (UID: \"eaba04e4-bc4a-41a0-aa79-631669451754\") " pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.046556 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eaba04e4-bc4a-41a0-aa79-631669451754-config-data-custom\") pod \"barbican-worker-9f949dfbc-7rr98\" (UID: \"eaba04e4-bc4a-41a0-aa79-631669451754\") " pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.046683 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jqkg\" (UniqueName: \"kubernetes.io/projected/eaba04e4-bc4a-41a0-aa79-631669451754-kube-api-access-2jqkg\") pod \"barbican-worker-9f949dfbc-7rr98\" (UID: \"eaba04e4-bc4a-41a0-aa79-631669451754\") " pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.046782 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaba04e4-bc4a-41a0-aa79-631669451754-logs\") pod \"barbican-worker-9f949dfbc-7rr98\" (UID: \"eaba04e4-bc4a-41a0-aa79-631669451754\") " pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.049037 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-869f779d85-d6lxp"] Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.050554 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.074806 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-d6lxp"] Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.109248 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-c8d5f4d66-7z5sv"] Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.115107 4646 generic.go:334] "Generic (PLEG): container finished" podID="7715c927-8035-48d3-908e-4f3838f296aa" containerID="0f87fafd1a0ea0968db590d01faad64a686fc1d57ebaa88420f6e38a55f65ba2" exitCode=0 Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.115152 4646 generic.go:334] "Generic (PLEG): container finished" podID="7715c927-8035-48d3-908e-4f3838f296aa" containerID="d9b3e1adfd5004498124ec7e0fc51643711a2c085b29b82477eac6993d0c4ee6" exitCode=2 Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.115623 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7715c927-8035-48d3-908e-4f3838f296aa","Type":"ContainerDied","Data":"0f87fafd1a0ea0968db590d01faad64a686fc1d57ebaa88420f6e38a55f65ba2"} Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.115654 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c8d5f4d66-7z5sv"] Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.115669 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7715c927-8035-48d3-908e-4f3838f296aa","Type":"ContainerDied","Data":"d9b3e1adfd5004498124ec7e0fc51643711a2c085b29b82477eac6993d0c4ee6"} Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.115741 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.125390 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.149245 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10ed882e-df14-4619-9a57-022edbefa36c-logs\") pod \"barbican-keystone-listener-6b6bfd6548-5hldq\" (UID: \"10ed882e-df14-4619-9a57-022edbefa36c\") " pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.149306 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jqkg\" (UniqueName: \"kubernetes.io/projected/eaba04e4-bc4a-41a0-aa79-631669451754-kube-api-access-2jqkg\") pod \"barbican-worker-9f949dfbc-7rr98\" (UID: \"eaba04e4-bc4a-41a0-aa79-631669451754\") " pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.149337 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-config\") pod \"dnsmasq-dns-869f779d85-d6lxp\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.149361 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-dns-svc\") pod \"dnsmasq-dns-869f779d85-d6lxp\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.149402 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaba04e4-bc4a-41a0-aa79-631669451754-logs\") pod \"barbican-worker-9f949dfbc-7rr98\" (UID: \"eaba04e4-bc4a-41a0-aa79-631669451754\") " pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.149438 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv9rv\" (UniqueName: \"kubernetes.io/projected/10ed882e-df14-4619-9a57-022edbefa36c-kube-api-access-lv9rv\") pod \"barbican-keystone-listener-6b6bfd6548-5hldq\" (UID: \"10ed882e-df14-4619-9a57-022edbefa36c\") " pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.149456 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-ovsdbserver-nb\") pod \"dnsmasq-dns-869f779d85-d6lxp\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.149511 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaba04e4-bc4a-41a0-aa79-631669451754-combined-ca-bundle\") pod \"barbican-worker-9f949dfbc-7rr98\" (UID: \"eaba04e4-bc4a-41a0-aa79-631669451754\") " pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.149561 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10ed882e-df14-4619-9a57-022edbefa36c-combined-ca-bundle\") pod \"barbican-keystone-listener-6b6bfd6548-5hldq\" (UID: \"10ed882e-df14-4619-9a57-022edbefa36c\") " pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.149589 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10ed882e-df14-4619-9a57-022edbefa36c-config-data-custom\") pod \"barbican-keystone-listener-6b6bfd6548-5hldq\" (UID: \"10ed882e-df14-4619-9a57-022edbefa36c\") " pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.149608 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktqcr\" (UniqueName: \"kubernetes.io/projected/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-kube-api-access-ktqcr\") pod \"dnsmasq-dns-869f779d85-d6lxp\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.149627 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-ovsdbserver-sb\") pod \"dnsmasq-dns-869f779d85-d6lxp\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.149652 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaba04e4-bc4a-41a0-aa79-631669451754-config-data\") pod \"barbican-worker-9f949dfbc-7rr98\" (UID: \"eaba04e4-bc4a-41a0-aa79-631669451754\") " pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.149674 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10ed882e-df14-4619-9a57-022edbefa36c-config-data\") pod \"barbican-keystone-listener-6b6bfd6548-5hldq\" (UID: \"10ed882e-df14-4619-9a57-022edbefa36c\") " pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.149702 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eaba04e4-bc4a-41a0-aa79-631669451754-config-data-custom\") pod \"barbican-worker-9f949dfbc-7rr98\" (UID: \"eaba04e4-bc4a-41a0-aa79-631669451754\") " pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.150817 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaba04e4-bc4a-41a0-aa79-631669451754-logs\") pod \"barbican-worker-9f949dfbc-7rr98\" (UID: \"eaba04e4-bc4a-41a0-aa79-631669451754\") " pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.160330 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaba04e4-bc4a-41a0-aa79-631669451754-combined-ca-bundle\") pod \"barbican-worker-9f949dfbc-7rr98\" (UID: \"eaba04e4-bc4a-41a0-aa79-631669451754\") " pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.169422 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaba04e4-bc4a-41a0-aa79-631669451754-config-data\") pod \"barbican-worker-9f949dfbc-7rr98\" (UID: \"eaba04e4-bc4a-41a0-aa79-631669451754\") " pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.177566 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eaba04e4-bc4a-41a0-aa79-631669451754-config-data-custom\") pod \"barbican-worker-9f949dfbc-7rr98\" (UID: \"eaba04e4-bc4a-41a0-aa79-631669451754\") " pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.194529 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jqkg\" (UniqueName: \"kubernetes.io/projected/eaba04e4-bc4a-41a0-aa79-631669451754-kube-api-access-2jqkg\") pod \"barbican-worker-9f949dfbc-7rr98\" (UID: \"eaba04e4-bc4a-41a0-aa79-631669451754\") " pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.250801 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10ed882e-df14-4619-9a57-022edbefa36c-combined-ca-bundle\") pod \"barbican-keystone-listener-6b6bfd6548-5hldq\" (UID: \"10ed882e-df14-4619-9a57-022edbefa36c\") " pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.250856 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10ed882e-df14-4619-9a57-022edbefa36c-config-data-custom\") pod \"barbican-keystone-listener-6b6bfd6548-5hldq\" (UID: \"10ed882e-df14-4619-9a57-022edbefa36c\") " pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.250887 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktqcr\" (UniqueName: \"kubernetes.io/projected/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-kube-api-access-ktqcr\") pod \"dnsmasq-dns-869f779d85-d6lxp\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.250920 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-ovsdbserver-sb\") pod \"dnsmasq-dns-869f779d85-d6lxp\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.250947 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10ed882e-df14-4619-9a57-022edbefa36c-config-data\") pod \"barbican-keystone-listener-6b6bfd6548-5hldq\" (UID: \"10ed882e-df14-4619-9a57-022edbefa36c\") " pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.251015 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10ed882e-df14-4619-9a57-022edbefa36c-logs\") pod \"barbican-keystone-listener-6b6bfd6548-5hldq\" (UID: \"10ed882e-df14-4619-9a57-022edbefa36c\") " pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.251051 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-config-data\") pod \"barbican-api-c8d5f4d66-7z5sv\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.251073 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-combined-ca-bundle\") pod \"barbican-api-c8d5f4d66-7z5sv\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.251109 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-config\") pod \"dnsmasq-dns-869f779d85-d6lxp\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.251137 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-dns-svc\") pod \"dnsmasq-dns-869f779d85-d6lxp\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.251196 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv9rv\" (UniqueName: \"kubernetes.io/projected/10ed882e-df14-4619-9a57-022edbefa36c-kube-api-access-lv9rv\") pod \"barbican-keystone-listener-6b6bfd6548-5hldq\" (UID: \"10ed882e-df14-4619-9a57-022edbefa36c\") " pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.251223 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-ovsdbserver-nb\") pod \"dnsmasq-dns-869f779d85-d6lxp\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.251251 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srb85\" (UniqueName: \"kubernetes.io/projected/b3d95cd5-dac1-4632-bb55-74af55174c0b-kube-api-access-srb85\") pod \"barbican-api-c8d5f4d66-7z5sv\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.251313 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-config-data-custom\") pod \"barbican-api-c8d5f4d66-7z5sv\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.251336 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3d95cd5-dac1-4632-bb55-74af55174c0b-logs\") pod \"barbican-api-c8d5f4d66-7z5sv\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.254013 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10ed882e-df14-4619-9a57-022edbefa36c-logs\") pod \"barbican-keystone-listener-6b6bfd6548-5hldq\" (UID: \"10ed882e-df14-4619-9a57-022edbefa36c\") " pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.255058 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-dns-svc\") pod \"dnsmasq-dns-869f779d85-d6lxp\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.255064 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-ovsdbserver-sb\") pod \"dnsmasq-dns-869f779d85-d6lxp\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.255068 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-config\") pod \"dnsmasq-dns-869f779d85-d6lxp\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.255643 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-ovsdbserver-nb\") pod \"dnsmasq-dns-869f779d85-d6lxp\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.255795 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10ed882e-df14-4619-9a57-022edbefa36c-config-data\") pod \"barbican-keystone-listener-6b6bfd6548-5hldq\" (UID: \"10ed882e-df14-4619-9a57-022edbefa36c\") " pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.258834 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10ed882e-df14-4619-9a57-022edbefa36c-config-data-custom\") pod \"barbican-keystone-listener-6b6bfd6548-5hldq\" (UID: \"10ed882e-df14-4619-9a57-022edbefa36c\") " pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.263638 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10ed882e-df14-4619-9a57-022edbefa36c-combined-ca-bundle\") pod \"barbican-keystone-listener-6b6bfd6548-5hldq\" (UID: \"10ed882e-df14-4619-9a57-022edbefa36c\") " pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.279916 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-9f949dfbc-7rr98" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.301125 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktqcr\" (UniqueName: \"kubernetes.io/projected/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-kube-api-access-ktqcr\") pod \"dnsmasq-dns-869f779d85-d6lxp\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.307089 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv9rv\" (UniqueName: \"kubernetes.io/projected/10ed882e-df14-4619-9a57-022edbefa36c-kube-api-access-lv9rv\") pod \"barbican-keystone-listener-6b6bfd6548-5hldq\" (UID: \"10ed882e-df14-4619-9a57-022edbefa36c\") " pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.325284 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.352452 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-config-data-custom\") pod \"barbican-api-c8d5f4d66-7z5sv\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.352526 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3d95cd5-dac1-4632-bb55-74af55174c0b-logs\") pod \"barbican-api-c8d5f4d66-7z5sv\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.352627 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-config-data\") pod \"barbican-api-c8d5f4d66-7z5sv\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.352647 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-combined-ca-bundle\") pod \"barbican-api-c8d5f4d66-7z5sv\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.352722 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srb85\" (UniqueName: \"kubernetes.io/projected/b3d95cd5-dac1-4632-bb55-74af55174c0b-kube-api-access-srb85\") pod \"barbican-api-c8d5f4d66-7z5sv\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.353192 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3d95cd5-dac1-4632-bb55-74af55174c0b-logs\") pod \"barbican-api-c8d5f4d66-7z5sv\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.356895 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-config-data-custom\") pod \"barbican-api-c8d5f4d66-7z5sv\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.366134 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-combined-ca-bundle\") pod \"barbican-api-c8d5f4d66-7z5sv\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.367174 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-config-data\") pod \"barbican-api-c8d5f4d66-7z5sv\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.384724 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.388276 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srb85\" (UniqueName: \"kubernetes.io/projected/b3d95cd5-dac1-4632-bb55-74af55174c0b-kube-api-access-srb85\") pod \"barbican-api-c8d5f4d66-7z5sv\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.629762 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:21 crc kubenswrapper[4646]: I1203 11:13:21.799079 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:21.868240 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-ovsdbserver-nb\") pod \"9aca34cf-ca0a-4449-9188-04d05766a266\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:21.868397 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-dns-svc\") pod \"9aca34cf-ca0a-4449-9188-04d05766a266\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:21.868433 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-config\") pod \"9aca34cf-ca0a-4449-9188-04d05766a266\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:21.868466 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-ovsdbserver-sb\") pod \"9aca34cf-ca0a-4449-9188-04d05766a266\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:21.868492 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz525\" (UniqueName: \"kubernetes.io/projected/9aca34cf-ca0a-4449-9188-04d05766a266-kube-api-access-sz525\") pod \"9aca34cf-ca0a-4449-9188-04d05766a266\" (UID: \"9aca34cf-ca0a-4449-9188-04d05766a266\") " Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:21.881555 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aca34cf-ca0a-4449-9188-04d05766a266-kube-api-access-sz525" (OuterVolumeSpecName: "kube-api-access-sz525") pod "9aca34cf-ca0a-4449-9188-04d05766a266" (UID: "9aca34cf-ca0a-4449-9188-04d05766a266"). InnerVolumeSpecName "kube-api-access-sz525". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:21.935632 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-config" (OuterVolumeSpecName: "config") pod "9aca34cf-ca0a-4449-9188-04d05766a266" (UID: "9aca34cf-ca0a-4449-9188-04d05766a266"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:21.970936 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:21.970960 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sz525\" (UniqueName: \"kubernetes.io/projected/9aca34cf-ca0a-4449-9188-04d05766a266-kube-api-access-sz525\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:21.974696 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9aca34cf-ca0a-4449-9188-04d05766a266" (UID: "9aca34cf-ca0a-4449-9188-04d05766a266"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:21.979760 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9aca34cf-ca0a-4449-9188-04d05766a266" (UID: "9aca34cf-ca0a-4449-9188-04d05766a266"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:21.981554 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9aca34cf-ca0a-4449-9188-04d05766a266" (UID: "9aca34cf-ca0a-4449-9188-04d05766a266"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.072354 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.072398 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.072408 4646 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aca34cf-ca0a-4449-9188-04d05766a266-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.123136 4646 generic.go:334] "Generic (PLEG): container finished" podID="9aca34cf-ca0a-4449-9188-04d05766a266" containerID="9ecdf4d0d834d19420e14a6d02d8b43d5740ef4470fd5322a49d6828c6807b53" exitCode=0 Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.123180 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" event={"ID":"9aca34cf-ca0a-4449-9188-04d05766a266","Type":"ContainerDied","Data":"9ecdf4d0d834d19420e14a6d02d8b43d5740ef4470fd5322a49d6828c6807b53"} Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.123206 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" event={"ID":"9aca34cf-ca0a-4449-9188-04d05766a266","Type":"ContainerDied","Data":"8c6ad3f743df7923262c547b799ae6a01f3939d0fad68f7a0db8df911cfad517"} Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.123222 4646 scope.go:117] "RemoveContainer" containerID="9ecdf4d0d834d19420e14a6d02d8b43d5740ef4470fd5322a49d6828c6807b53" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.123377 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-7j5cd" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.181513 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-7j5cd"] Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.186812 4646 scope.go:117] "RemoveContainer" containerID="6645a7892ca5cacbea98ff06dcf01bce92f9e0bcbde52602865e8ba4ba824745" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.189771 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-7j5cd"] Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.196947 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-9f949dfbc-7rr98"] Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.210150 4646 scope.go:117] "RemoveContainer" containerID="9ecdf4d0d834d19420e14a6d02d8b43d5740ef4470fd5322a49d6828c6807b53" Dec 03 11:13:22 crc kubenswrapper[4646]: E1203 11:13:22.211441 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ecdf4d0d834d19420e14a6d02d8b43d5740ef4470fd5322a49d6828c6807b53\": container with ID starting with 9ecdf4d0d834d19420e14a6d02d8b43d5740ef4470fd5322a49d6828c6807b53 not found: ID does not exist" containerID="9ecdf4d0d834d19420e14a6d02d8b43d5740ef4470fd5322a49d6828c6807b53" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.211469 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ecdf4d0d834d19420e14a6d02d8b43d5740ef4470fd5322a49d6828c6807b53"} err="failed to get container status \"9ecdf4d0d834d19420e14a6d02d8b43d5740ef4470fd5322a49d6828c6807b53\": rpc error: code = NotFound desc = could not find container \"9ecdf4d0d834d19420e14a6d02d8b43d5740ef4470fd5322a49d6828c6807b53\": container with ID starting with 9ecdf4d0d834d19420e14a6d02d8b43d5740ef4470fd5322a49d6828c6807b53 not found: ID does not exist" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.211491 4646 scope.go:117] "RemoveContainer" containerID="6645a7892ca5cacbea98ff06dcf01bce92f9e0bcbde52602865e8ba4ba824745" Dec 03 11:13:22 crc kubenswrapper[4646]: E1203 11:13:22.212311 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6645a7892ca5cacbea98ff06dcf01bce92f9e0bcbde52602865e8ba4ba824745\": container with ID starting with 6645a7892ca5cacbea98ff06dcf01bce92f9e0bcbde52602865e8ba4ba824745 not found: ID does not exist" containerID="6645a7892ca5cacbea98ff06dcf01bce92f9e0bcbde52602865e8ba4ba824745" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.212367 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6645a7892ca5cacbea98ff06dcf01bce92f9e0bcbde52602865e8ba4ba824745"} err="failed to get container status \"6645a7892ca5cacbea98ff06dcf01bce92f9e0bcbde52602865e8ba4ba824745\": rpc error: code = NotFound desc = could not find container \"6645a7892ca5cacbea98ff06dcf01bce92f9e0bcbde52602865e8ba4ba824745\": container with ID starting with 6645a7892ca5cacbea98ff06dcf01bce92f9e0bcbde52602865e8ba4ba824745 not found: ID does not exist" Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.853494 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6b6bfd6548-5hldq"] Dec 03 11:13:22 crc kubenswrapper[4646]: W1203 11:13:22.861265 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10ed882e_df14_4619_9a57_022edbefa36c.slice/crio-04e3496ec7a9d50da984e876513816f3866bf0f17566ddc6823d2160461c1c82 WatchSource:0}: Error finding container 04e3496ec7a9d50da984e876513816f3866bf0f17566ddc6823d2160461c1c82: Status 404 returned error can't find the container with id 04e3496ec7a9d50da984e876513816f3866bf0f17566ddc6823d2160461c1c82 Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.868811 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-d6lxp"] Dec 03 11:13:22 crc kubenswrapper[4646]: I1203 11:13:22.991382 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c8d5f4d66-7z5sv"] Dec 03 11:13:22 crc kubenswrapper[4646]: W1203 11:13:22.993443 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3d95cd5_dac1_4632_bb55_74af55174c0b.slice/crio-2067f23bb6007d03d172f339facf799c74e7af51cb98f73b509d0186d9235432 WatchSource:0}: Error finding container 2067f23bb6007d03d172f339facf799c74e7af51cb98f73b509d0186d9235432: Status 404 returned error can't find the container with id 2067f23bb6007d03d172f339facf799c74e7af51cb98f73b509d0186d9235432 Dec 03 11:13:23 crc kubenswrapper[4646]: I1203 11:13:23.144676 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" event={"ID":"10ed882e-df14-4619-9a57-022edbefa36c","Type":"ContainerStarted","Data":"04e3496ec7a9d50da984e876513816f3866bf0f17566ddc6823d2160461c1c82"} Dec 03 11:13:23 crc kubenswrapper[4646]: I1203 11:13:23.147185 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-9f949dfbc-7rr98" event={"ID":"eaba04e4-bc4a-41a0-aa79-631669451754","Type":"ContainerStarted","Data":"76f7db37aee85e38a7489982446163da75c63e9c4d54e93d8c811b71fd430fc2"} Dec 03 11:13:23 crc kubenswrapper[4646]: I1203 11:13:23.148827 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c8d5f4d66-7z5sv" event={"ID":"b3d95cd5-dac1-4632-bb55-74af55174c0b","Type":"ContainerStarted","Data":"2067f23bb6007d03d172f339facf799c74e7af51cb98f73b509d0186d9235432"} Dec 03 11:13:23 crc kubenswrapper[4646]: I1203 11:13:23.156649 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-d6lxp" event={"ID":"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005","Type":"ContainerStarted","Data":"bde7d201205a24673a3153a1ecd13230176ae3bfac1bfe5cde58e9241d4f2437"} Dec 03 11:13:23 crc kubenswrapper[4646]: I1203 11:13:23.156712 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-d6lxp" event={"ID":"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005","Type":"ContainerStarted","Data":"8060fd6ef9343dadf328586c9faa2e712725728bf00bc6d3e58e77e0ae86007d"} Dec 03 11:13:23 crc kubenswrapper[4646]: I1203 11:13:23.863878 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aca34cf-ca0a-4449-9188-04d05766a266" path="/var/lib/kubelet/pods/9aca34cf-ca0a-4449-9188-04d05766a266/volumes" Dec 03 11:13:24 crc kubenswrapper[4646]: I1203 11:13:24.166196 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c8d5f4d66-7z5sv" event={"ID":"b3d95cd5-dac1-4632-bb55-74af55174c0b","Type":"ContainerStarted","Data":"043dab58c03b248330030af1fe35f42b1c39d73c3d673b184e94d224adb8347c"} Dec 03 11:13:24 crc kubenswrapper[4646]: I1203 11:13:24.166270 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c8d5f4d66-7z5sv" event={"ID":"b3d95cd5-dac1-4632-bb55-74af55174c0b","Type":"ContainerStarted","Data":"b43f7e2e0255482a5fd87e63eee44cd517625e027c736cace5068dea0816d86b"} Dec 03 11:13:24 crc kubenswrapper[4646]: I1203 11:13:24.166641 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:24 crc kubenswrapper[4646]: I1203 11:13:24.166685 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:24 crc kubenswrapper[4646]: I1203 11:13:24.168148 4646 generic.go:334] "Generic (PLEG): container finished" podID="6fb3b18d-41af-427b-b2b1-e3b8cbbb0005" containerID="bde7d201205a24673a3153a1ecd13230176ae3bfac1bfe5cde58e9241d4f2437" exitCode=0 Dec 03 11:13:24 crc kubenswrapper[4646]: I1203 11:13:24.168418 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-d6lxp" event={"ID":"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005","Type":"ContainerDied","Data":"bde7d201205a24673a3153a1ecd13230176ae3bfac1bfe5cde58e9241d4f2437"} Dec 03 11:13:24 crc kubenswrapper[4646]: I1203 11:13:24.173559 4646 generic.go:334] "Generic (PLEG): container finished" podID="7715c927-8035-48d3-908e-4f3838f296aa" containerID="487ee5827c75a4af2682b7272eaee399afa12a6e73d723eae253df638228a6af" exitCode=0 Dec 03 11:13:24 crc kubenswrapper[4646]: I1203 11:13:24.173612 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7715c927-8035-48d3-908e-4f3838f296aa","Type":"ContainerDied","Data":"487ee5827c75a4af2682b7272eaee399afa12a6e73d723eae253df638228a6af"} Dec 03 11:13:24 crc kubenswrapper[4646]: I1203 11:13:24.192957 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-c8d5f4d66-7z5sv" podStartSLOduration=3.192929998 podStartE2EDuration="3.192929998s" podCreationTimestamp="2025-12-03 11:13:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:13:24.186110732 +0000 UTC m=+1180.649166867" watchObservedRunningTime="2025-12-03 11:13:24.192929998 +0000 UTC m=+1180.655986133" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.137579 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.142672 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6d775b8f86-64hmf"] Dec 03 11:13:25 crc kubenswrapper[4646]: E1203 11:13:25.146064 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7715c927-8035-48d3-908e-4f3838f296aa" containerName="sg-core" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.146104 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="7715c927-8035-48d3-908e-4f3838f296aa" containerName="sg-core" Dec 03 11:13:25 crc kubenswrapper[4646]: E1203 11:13:25.146120 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7715c927-8035-48d3-908e-4f3838f296aa" containerName="proxy-httpd" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.146126 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="7715c927-8035-48d3-908e-4f3838f296aa" containerName="proxy-httpd" Dec 03 11:13:25 crc kubenswrapper[4646]: E1203 11:13:25.146165 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7715c927-8035-48d3-908e-4f3838f296aa" containerName="ceilometer-notification-agent" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.146172 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="7715c927-8035-48d3-908e-4f3838f296aa" containerName="ceilometer-notification-agent" Dec 03 11:13:25 crc kubenswrapper[4646]: E1203 11:13:25.146184 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aca34cf-ca0a-4449-9188-04d05766a266" containerName="init" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.146189 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aca34cf-ca0a-4449-9188-04d05766a266" containerName="init" Dec 03 11:13:25 crc kubenswrapper[4646]: E1203 11:13:25.146206 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aca34cf-ca0a-4449-9188-04d05766a266" containerName="dnsmasq-dns" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.146212 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aca34cf-ca0a-4449-9188-04d05766a266" containerName="dnsmasq-dns" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.146494 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="7715c927-8035-48d3-908e-4f3838f296aa" containerName="sg-core" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.146515 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="7715c927-8035-48d3-908e-4f3838f296aa" containerName="ceilometer-notification-agent" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.146530 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aca34cf-ca0a-4449-9188-04d05766a266" containerName="dnsmasq-dns" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.146554 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="7715c927-8035-48d3-908e-4f3838f296aa" containerName="proxy-httpd" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.147488 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.173283 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.173604 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.180417 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6d775b8f86-64hmf"] Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.203971 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" event={"ID":"10ed882e-df14-4619-9a57-022edbefa36c","Type":"ContainerStarted","Data":"bcc1dca372c7056ff50da88154af2935930064295cbe67272c6848cc5dbaf960"} Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.205000 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-9f949dfbc-7rr98" event={"ID":"eaba04e4-bc4a-41a0-aa79-631669451754","Type":"ContainerStarted","Data":"2cfdc3542663d398d1d7f6043c4519037924e02467fb40a3a02fd83b6a2b856c"} Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.224095 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-d6lxp" event={"ID":"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005","Type":"ContainerStarted","Data":"276ca750a898323d5ccda9ed35f43e91737a15239431d4b732e87c7f43128783"} Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.225084 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.230770 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.231183 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7715c927-8035-48d3-908e-4f3838f296aa","Type":"ContainerDied","Data":"6da83e9f1049a4caafa2d3961db0d5afb1c058ebdfd632ce1b7f72842ce5a704"} Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.231216 4646 scope.go:117] "RemoveContainer" containerID="0f87fafd1a0ea0968db590d01faad64a686fc1d57ebaa88420f6e38a55f65ba2" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.249692 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-combined-ca-bundle\") pod \"7715c927-8035-48d3-908e-4f3838f296aa\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.249787 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-sg-core-conf-yaml\") pod \"7715c927-8035-48d3-908e-4f3838f296aa\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.249915 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-config-data\") pod \"7715c927-8035-48d3-908e-4f3838f296aa\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.249937 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7715c927-8035-48d3-908e-4f3838f296aa-log-httpd\") pod \"7715c927-8035-48d3-908e-4f3838f296aa\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.249961 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cx9n\" (UniqueName: \"kubernetes.io/projected/7715c927-8035-48d3-908e-4f3838f296aa-kube-api-access-2cx9n\") pod \"7715c927-8035-48d3-908e-4f3838f296aa\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.249979 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-scripts\") pod \"7715c927-8035-48d3-908e-4f3838f296aa\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.250017 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7715c927-8035-48d3-908e-4f3838f296aa-run-httpd\") pod \"7715c927-8035-48d3-908e-4f3838f296aa\" (UID: \"7715c927-8035-48d3-908e-4f3838f296aa\") " Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.250184 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07721619-ea6b-4cb8-95bb-f5801fc848e3-internal-tls-certs\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.250212 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07721619-ea6b-4cb8-95bb-f5801fc848e3-combined-ca-bundle\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.250232 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07721619-ea6b-4cb8-95bb-f5801fc848e3-config-data\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.250250 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07721619-ea6b-4cb8-95bb-f5801fc848e3-logs\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.250302 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5qxz\" (UniqueName: \"kubernetes.io/projected/07721619-ea6b-4cb8-95bb-f5801fc848e3-kube-api-access-r5qxz\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.250408 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07721619-ea6b-4cb8-95bb-f5801fc848e3-public-tls-certs\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.250433 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07721619-ea6b-4cb8-95bb-f5801fc848e3-config-data-custom\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.250962 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-869f779d85-d6lxp" podStartSLOduration=5.250939985 podStartE2EDuration="5.250939985s" podCreationTimestamp="2025-12-03 11:13:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:13:25.244791017 +0000 UTC m=+1181.707847162" watchObservedRunningTime="2025-12-03 11:13:25.250939985 +0000 UTC m=+1181.713996110" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.251985 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7715c927-8035-48d3-908e-4f3838f296aa-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7715c927-8035-48d3-908e-4f3838f296aa" (UID: "7715c927-8035-48d3-908e-4f3838f296aa"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.252748 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7715c927-8035-48d3-908e-4f3838f296aa-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7715c927-8035-48d3-908e-4f3838f296aa" (UID: "7715c927-8035-48d3-908e-4f3838f296aa"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.260127 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7715c927-8035-48d3-908e-4f3838f296aa-kube-api-access-2cx9n" (OuterVolumeSpecName: "kube-api-access-2cx9n") pod "7715c927-8035-48d3-908e-4f3838f296aa" (UID: "7715c927-8035-48d3-908e-4f3838f296aa"). InnerVolumeSpecName "kube-api-access-2cx9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.266448 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-scripts" (OuterVolumeSpecName: "scripts") pod "7715c927-8035-48d3-908e-4f3838f296aa" (UID: "7715c927-8035-48d3-908e-4f3838f296aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.288651 4646 scope.go:117] "RemoveContainer" containerID="d9b3e1adfd5004498124ec7e0fc51643711a2c085b29b82477eac6993d0c4ee6" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.313279 4646 scope.go:117] "RemoveContainer" containerID="487ee5827c75a4af2682b7272eaee399afa12a6e73d723eae253df638228a6af" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.315051 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7715c927-8035-48d3-908e-4f3838f296aa" (UID: "7715c927-8035-48d3-908e-4f3838f296aa"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.315134 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7715c927-8035-48d3-908e-4f3838f296aa" (UID: "7715c927-8035-48d3-908e-4f3838f296aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.352102 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5qxz\" (UniqueName: \"kubernetes.io/projected/07721619-ea6b-4cb8-95bb-f5801fc848e3-kube-api-access-r5qxz\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.352258 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07721619-ea6b-4cb8-95bb-f5801fc848e3-public-tls-certs\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.352275 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07721619-ea6b-4cb8-95bb-f5801fc848e3-config-data-custom\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.352295 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07721619-ea6b-4cb8-95bb-f5801fc848e3-internal-tls-certs\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.352312 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07721619-ea6b-4cb8-95bb-f5801fc848e3-combined-ca-bundle\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.352326 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07721619-ea6b-4cb8-95bb-f5801fc848e3-config-data\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.352344 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07721619-ea6b-4cb8-95bb-f5801fc848e3-logs\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.352910 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07721619-ea6b-4cb8-95bb-f5801fc848e3-logs\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.355216 4646 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7715c927-8035-48d3-908e-4f3838f296aa-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.355235 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cx9n\" (UniqueName: \"kubernetes.io/projected/7715c927-8035-48d3-908e-4f3838f296aa-kube-api-access-2cx9n\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.355244 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.355253 4646 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7715c927-8035-48d3-908e-4f3838f296aa-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.355262 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.355272 4646 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.359244 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07721619-ea6b-4cb8-95bb-f5801fc848e3-internal-tls-certs\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.360683 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07721619-ea6b-4cb8-95bb-f5801fc848e3-public-tls-certs\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.360816 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-config-data" (OuterVolumeSpecName: "config-data") pod "7715c927-8035-48d3-908e-4f3838f296aa" (UID: "7715c927-8035-48d3-908e-4f3838f296aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.361847 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07721619-ea6b-4cb8-95bb-f5801fc848e3-combined-ca-bundle\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.362008 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07721619-ea6b-4cb8-95bb-f5801fc848e3-config-data-custom\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.363936 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07721619-ea6b-4cb8-95bb-f5801fc848e3-config-data\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.370968 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5qxz\" (UniqueName: \"kubernetes.io/projected/07721619-ea6b-4cb8-95bb-f5801fc848e3-kube-api-access-r5qxz\") pod \"barbican-api-6d775b8f86-64hmf\" (UID: \"07721619-ea6b-4cb8-95bb-f5801fc848e3\") " pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.457170 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7715c927-8035-48d3-908e-4f3838f296aa-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.496737 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.618216 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.644648 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.696054 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.698324 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.702236 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.706264 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.709655 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.864446 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-config-data\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.864486 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c9f14af-75f4-4b2a-97fc-97aa12438933-log-httpd\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.864521 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c9f14af-75f4-4b2a-97fc-97aa12438933-run-httpd\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.864559 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.864611 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-scripts\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.864625 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.864691 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46btx\" (UniqueName: \"kubernetes.io/projected/6c9f14af-75f4-4b2a-97fc-97aa12438933-kube-api-access-46btx\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.895700 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7715c927-8035-48d3-908e-4f3838f296aa" path="/var/lib/kubelet/pods/7715c927-8035-48d3-908e-4f3838f296aa/volumes" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.966789 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46btx\" (UniqueName: \"kubernetes.io/projected/6c9f14af-75f4-4b2a-97fc-97aa12438933-kube-api-access-46btx\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.966865 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-config-data\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.966885 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c9f14af-75f4-4b2a-97fc-97aa12438933-log-httpd\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.966918 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c9f14af-75f4-4b2a-97fc-97aa12438933-run-httpd\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.966970 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.967068 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-scripts\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.967090 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.967915 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c9f14af-75f4-4b2a-97fc-97aa12438933-run-httpd\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.968581 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c9f14af-75f4-4b2a-97fc-97aa12438933-log-httpd\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.982727 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-config-data\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.984809 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.985018 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46btx\" (UniqueName: \"kubernetes.io/projected/6c9f14af-75f4-4b2a-97fc-97aa12438933-kube-api-access-46btx\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.985131 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:25 crc kubenswrapper[4646]: I1203 11:13:25.987065 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-scripts\") pod \"ceilometer-0\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " pod="openstack/ceilometer-0" Dec 03 11:13:26 crc kubenswrapper[4646]: I1203 11:13:26.048165 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6d775b8f86-64hmf"] Dec 03 11:13:26 crc kubenswrapper[4646]: I1203 11:13:26.060565 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:13:26 crc kubenswrapper[4646]: I1203 11:13:26.268964 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" event={"ID":"10ed882e-df14-4619-9a57-022edbefa36c","Type":"ContainerStarted","Data":"c148db36528eae1ec2647b2a373d296568cc4c2864ad023d3ef4c6501fb4fca0"} Dec 03 11:13:26 crc kubenswrapper[4646]: I1203 11:13:26.280298 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-9f949dfbc-7rr98" event={"ID":"eaba04e4-bc4a-41a0-aa79-631669451754","Type":"ContainerStarted","Data":"4154643aab7b856433c0101a17c84d8e79f362720c1368d593e1a0fff6de7300"} Dec 03 11:13:26 crc kubenswrapper[4646]: I1203 11:13:26.292324 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d775b8f86-64hmf" event={"ID":"07721619-ea6b-4cb8-95bb-f5801fc848e3","Type":"ContainerStarted","Data":"f20d3f701b393f506109e1883f3cefa38fdf6f3cf70ff44dc2ae07d1fed3a682"} Dec 03 11:13:26 crc kubenswrapper[4646]: I1203 11:13:26.313294 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6b6bfd6548-5hldq" podStartSLOduration=4.450192603 podStartE2EDuration="6.31326742s" podCreationTimestamp="2025-12-03 11:13:20 +0000 UTC" firstStartedPulling="2025-12-03 11:13:22.879081052 +0000 UTC m=+1179.342137187" lastFinishedPulling="2025-12-03 11:13:24.742155869 +0000 UTC m=+1181.205212004" observedRunningTime="2025-12-03 11:13:26.293436799 +0000 UTC m=+1182.756492934" watchObservedRunningTime="2025-12-03 11:13:26.31326742 +0000 UTC m=+1182.776323555" Dec 03 11:13:26 crc kubenswrapper[4646]: I1203 11:13:26.334873 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-9f949dfbc-7rr98" podStartSLOduration=3.804053316 podStartE2EDuration="6.33484987s" podCreationTimestamp="2025-12-03 11:13:20 +0000 UTC" firstStartedPulling="2025-12-03 11:13:22.2097157 +0000 UTC m=+1178.672771835" lastFinishedPulling="2025-12-03 11:13:24.740512254 +0000 UTC m=+1181.203568389" observedRunningTime="2025-12-03 11:13:26.323179891 +0000 UTC m=+1182.786236026" watchObservedRunningTime="2025-12-03 11:13:26.33484987 +0000 UTC m=+1182.797906005" Dec 03 11:13:26 crc kubenswrapper[4646]: I1203 11:13:26.556115 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:13:27 crc kubenswrapper[4646]: I1203 11:13:27.301833 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6c9f14af-75f4-4b2a-97fc-97aa12438933","Type":"ContainerStarted","Data":"248c45089ab456646945e0c6a14adceffa4cd008bd2d5dcc864add7fb295cb5f"} Dec 03 11:13:27 crc kubenswrapper[4646]: I1203 11:13:27.304588 4646 generic.go:334] "Generic (PLEG): container finished" podID="76443d6c-7930-4c62-abbb-d7ff831232ca" containerID="e7a283c6e0898451966968b2ef55a759b711684e8bdf653e930f30c52b97ed19" exitCode=0 Dec 03 11:13:27 crc kubenswrapper[4646]: I1203 11:13:27.304653 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rtwcp" event={"ID":"76443d6c-7930-4c62-abbb-d7ff831232ca","Type":"ContainerDied","Data":"e7a283c6e0898451966968b2ef55a759b711684e8bdf653e930f30c52b97ed19"} Dec 03 11:13:27 crc kubenswrapper[4646]: I1203 11:13:27.314446 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d775b8f86-64hmf" event={"ID":"07721619-ea6b-4cb8-95bb-f5801fc848e3","Type":"ContainerStarted","Data":"1a1bd576e46ca8944cbd1bae0a682aba0812b5dbe30fb8cb5eea6381af8a0c4f"} Dec 03 11:13:27 crc kubenswrapper[4646]: I1203 11:13:27.314728 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d775b8f86-64hmf" event={"ID":"07721619-ea6b-4cb8-95bb-f5801fc848e3","Type":"ContainerStarted","Data":"dd84b65afc119f9b110604d1732d880ca9f6ea117a961fdf80d8436d54090f8d"} Dec 03 11:13:27 crc kubenswrapper[4646]: I1203 11:13:27.348837 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6d775b8f86-64hmf" podStartSLOduration=2.348819825 podStartE2EDuration="2.348819825s" podCreationTimestamp="2025-12-03 11:13:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:13:27.347317644 +0000 UTC m=+1183.810373779" watchObservedRunningTime="2025-12-03 11:13:27.348819825 +0000 UTC m=+1183.811875960" Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.328593 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6c9f14af-75f4-4b2a-97fc-97aa12438933","Type":"ContainerStarted","Data":"e6968eaa9205bf5248c420c091e74ce4d23ab8f237df5357e8c27c18ea9d4f63"} Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.330079 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6c9f14af-75f4-4b2a-97fc-97aa12438933","Type":"ContainerStarted","Data":"2106226de82078f155cc8466d5d9925b0cde38c2ebe6413d587efe6acff2c697"} Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.331025 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.331143 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.662770 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.821935 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76443d6c-7930-4c62-abbb-d7ff831232ca-etc-machine-id\") pod \"76443d6c-7930-4c62-abbb-d7ff831232ca\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.822058 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/76443d6c-7930-4c62-abbb-d7ff831232ca-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "76443d6c-7930-4c62-abbb-d7ff831232ca" (UID: "76443d6c-7930-4c62-abbb-d7ff831232ca"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.822113 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzqhk\" (UniqueName: \"kubernetes.io/projected/76443d6c-7930-4c62-abbb-d7ff831232ca-kube-api-access-vzqhk\") pod \"76443d6c-7930-4c62-abbb-d7ff831232ca\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.822174 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-combined-ca-bundle\") pod \"76443d6c-7930-4c62-abbb-d7ff831232ca\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.822215 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-scripts\") pod \"76443d6c-7930-4c62-abbb-d7ff831232ca\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.822283 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-db-sync-config-data\") pod \"76443d6c-7930-4c62-abbb-d7ff831232ca\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.822318 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-config-data\") pod \"76443d6c-7930-4c62-abbb-d7ff831232ca\" (UID: \"76443d6c-7930-4c62-abbb-d7ff831232ca\") " Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.822871 4646 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76443d6c-7930-4c62-abbb-d7ff831232ca-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.827511 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-scripts" (OuterVolumeSpecName: "scripts") pod "76443d6c-7930-4c62-abbb-d7ff831232ca" (UID: "76443d6c-7930-4c62-abbb-d7ff831232ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.827646 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76443d6c-7930-4c62-abbb-d7ff831232ca-kube-api-access-vzqhk" (OuterVolumeSpecName: "kube-api-access-vzqhk") pod "76443d6c-7930-4c62-abbb-d7ff831232ca" (UID: "76443d6c-7930-4c62-abbb-d7ff831232ca"). InnerVolumeSpecName "kube-api-access-vzqhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.829564 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "76443d6c-7930-4c62-abbb-d7ff831232ca" (UID: "76443d6c-7930-4c62-abbb-d7ff831232ca"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.850594 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76443d6c-7930-4c62-abbb-d7ff831232ca" (UID: "76443d6c-7930-4c62-abbb-d7ff831232ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.883592 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-config-data" (OuterVolumeSpecName: "config-data") pod "76443d6c-7930-4c62-abbb-d7ff831232ca" (UID: "76443d6c-7930-4c62-abbb-d7ff831232ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.925265 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzqhk\" (UniqueName: \"kubernetes.io/projected/76443d6c-7930-4c62-abbb-d7ff831232ca-kube-api-access-vzqhk\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.925319 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.925362 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.925383 4646 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:28 crc kubenswrapper[4646]: I1203 11:13:28.925400 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76443d6c-7930-4c62-abbb-d7ff831232ca-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.342141 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6c9f14af-75f4-4b2a-97fc-97aa12438933","Type":"ContainerStarted","Data":"10b5c6cd346c8e48d288dc5e747ab8d805b55ee8e20a99e32d248e1b17f53bac"} Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.348202 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rtwcp" event={"ID":"76443d6c-7930-4c62-abbb-d7ff831232ca","Type":"ContainerDied","Data":"3b715aa20bee782c924fe115d7211eb838580ee2cee80a46adc742debcc50616"} Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.348449 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b715aa20bee782c924fe115d7211eb838580ee2cee80a46adc742debcc50616" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.348244 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rtwcp" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.627408 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 11:13:29 crc kubenswrapper[4646]: E1203 11:13:29.628011 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76443d6c-7930-4c62-abbb-d7ff831232ca" containerName="cinder-db-sync" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.628027 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="76443d6c-7930-4c62-abbb-d7ff831232ca" containerName="cinder-db-sync" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.628200 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="76443d6c-7930-4c62-abbb-d7ff831232ca" containerName="cinder-db-sync" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.629158 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.651481 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-cvt9x" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.654225 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.654770 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.655005 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.687414 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.701870 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-d6lxp"] Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.702288 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-869f779d85-d6lxp" podUID="6fb3b18d-41af-427b-b2b1-e3b8cbbb0005" containerName="dnsmasq-dns" containerID="cri-o://276ca750a898323d5ccda9ed35f43e91737a15239431d4b732e87c7f43128783" gracePeriod=10 Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.755249 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.755311 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.755411 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96k4l\" (UniqueName: \"kubernetes.io/projected/ac255daf-f97d-47c1-9cfa-436577e24c50-kube-api-access-96k4l\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.755466 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-config-data\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.755515 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-scripts\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.755584 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ac255daf-f97d-47c1-9cfa-436577e24c50-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.776036 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-ddhcl"] Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.787049 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.804184 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-ddhcl"] Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.857972 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ac255daf-f97d-47c1-9cfa-436577e24c50-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.858037 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.858073 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.858092 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96k4l\" (UniqueName: \"kubernetes.io/projected/ac255daf-f97d-47c1-9cfa-436577e24c50-kube-api-access-96k4l\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.858136 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-config-data\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.858162 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-scripts\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.862433 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ac255daf-f97d-47c1-9cfa-436577e24c50-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.871826 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-scripts\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.872315 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-config-data\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.872749 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.873198 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.932392 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96k4l\" (UniqueName: \"kubernetes.io/projected/ac255daf-f97d-47c1-9cfa-436577e24c50-kube-api-access-96k4l\") pod \"cinder-scheduler-0\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.946572 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.959242 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-dns-svc\") pod \"dnsmasq-dns-58db5546cc-ddhcl\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.959318 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-ovsdbserver-sb\") pod \"dnsmasq-dns-58db5546cc-ddhcl\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.959404 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-config\") pod \"dnsmasq-dns-58db5546cc-ddhcl\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.959494 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kpvs\" (UniqueName: \"kubernetes.io/projected/196f11c5-52de-42c3-ac45-dbaad2588b1d-kube-api-access-5kpvs\") pod \"dnsmasq-dns-58db5546cc-ddhcl\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:29 crc kubenswrapper[4646]: I1203 11:13:29.959528 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-ovsdbserver-nb\") pod \"dnsmasq-dns-58db5546cc-ddhcl\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.012008 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.013536 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.030702 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.036162 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.072580 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-dns-svc\") pod \"dnsmasq-dns-58db5546cc-ddhcl\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.072655 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-ovsdbserver-sb\") pod \"dnsmasq-dns-58db5546cc-ddhcl\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.072736 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-config\") pod \"dnsmasq-dns-58db5546cc-ddhcl\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.072788 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kpvs\" (UniqueName: \"kubernetes.io/projected/196f11c5-52de-42c3-ac45-dbaad2588b1d-kube-api-access-5kpvs\") pod \"dnsmasq-dns-58db5546cc-ddhcl\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.072826 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-ovsdbserver-nb\") pod \"dnsmasq-dns-58db5546cc-ddhcl\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.073616 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-ovsdbserver-sb\") pod \"dnsmasq-dns-58db5546cc-ddhcl\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.082945 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-dns-svc\") pod \"dnsmasq-dns-58db5546cc-ddhcl\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.085494 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-config\") pod \"dnsmasq-dns-58db5546cc-ddhcl\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.086583 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-ovsdbserver-nb\") pod \"dnsmasq-dns-58db5546cc-ddhcl\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.140524 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kpvs\" (UniqueName: \"kubernetes.io/projected/196f11c5-52de-42c3-ac45-dbaad2588b1d-kube-api-access-5kpvs\") pod \"dnsmasq-dns-58db5546cc-ddhcl\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.185258 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.186310 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a559e2e6-4fbc-4c51-99fb-5e971142d41e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.186343 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a559e2e6-4fbc-4c51-99fb-5e971142d41e-logs\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.186411 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-config-data-custom\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.186438 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bbqk\" (UniqueName: \"kubernetes.io/projected/a559e2e6-4fbc-4c51-99fb-5e971142d41e-kube-api-access-7bbqk\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.186466 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-scripts\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.186502 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-config-data\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.291533 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.291642 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a559e2e6-4fbc-4c51-99fb-5e971142d41e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.291658 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a559e2e6-4fbc-4c51-99fb-5e971142d41e-logs\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.291692 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-config-data-custom\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.291714 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bbqk\" (UniqueName: \"kubernetes.io/projected/a559e2e6-4fbc-4c51-99fb-5e971142d41e-kube-api-access-7bbqk\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.291743 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-scripts\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.291771 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-config-data\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.293820 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a559e2e6-4fbc-4c51-99fb-5e971142d41e-logs\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.297413 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a559e2e6-4fbc-4c51-99fb-5e971142d41e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.298632 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.354865 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.358499 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-config-data-custom\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.359148 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-config-data\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.362366 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-scripts\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.376928 4646 generic.go:334] "Generic (PLEG): container finished" podID="6fb3b18d-41af-427b-b2b1-e3b8cbbb0005" containerID="276ca750a898323d5ccda9ed35f43e91737a15239431d4b732e87c7f43128783" exitCode=0 Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.376969 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-d6lxp" event={"ID":"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005","Type":"ContainerDied","Data":"276ca750a898323d5ccda9ed35f43e91737a15239431d4b732e87c7f43128783"} Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.381239 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bbqk\" (UniqueName: \"kubernetes.io/projected/a559e2e6-4fbc-4c51-99fb-5e971142d41e-kube-api-access-7bbqk\") pod \"cinder-api-0\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.408354 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.592829 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.701525 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-config\") pod \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.701572 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-ovsdbserver-sb\") pod \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.701658 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-dns-svc\") pod \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.701766 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktqcr\" (UniqueName: \"kubernetes.io/projected/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-kube-api-access-ktqcr\") pod \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.701827 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-ovsdbserver-nb\") pod \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\" (UID: \"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005\") " Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.727300 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-kube-api-access-ktqcr" (OuterVolumeSpecName: "kube-api-access-ktqcr") pod "6fb3b18d-41af-427b-b2b1-e3b8cbbb0005" (UID: "6fb3b18d-41af-427b-b2b1-e3b8cbbb0005"). InnerVolumeSpecName "kube-api-access-ktqcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.805673 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktqcr\" (UniqueName: \"kubernetes.io/projected/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-kube-api-access-ktqcr\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.809189 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6fb3b18d-41af-427b-b2b1-e3b8cbbb0005" (UID: "6fb3b18d-41af-427b-b2b1-e3b8cbbb0005"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.912373 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.930636 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.965843 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6fb3b18d-41af-427b-b2b1-e3b8cbbb0005" (UID: "6fb3b18d-41af-427b-b2b1-e3b8cbbb0005"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.969955 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-config" (OuterVolumeSpecName: "config") pod "6fb3b18d-41af-427b-b2b1-e3b8cbbb0005" (UID: "6fb3b18d-41af-427b-b2b1-e3b8cbbb0005"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:30 crc kubenswrapper[4646]: I1203 11:13:30.984014 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-ddhcl"] Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.007868 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6fb3b18d-41af-427b-b2b1-e3b8cbbb0005" (UID: "6fb3b18d-41af-427b-b2b1-e3b8cbbb0005"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.031748 4646 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.031774 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.031785 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.106507 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.389779 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6c9f14af-75f4-4b2a-97fc-97aa12438933","Type":"ContainerStarted","Data":"d9f3e82110bbcce0aa356967fc9c8d8e8a9cae58f1a09379a288ecd647359094"} Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.390223 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.395979 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-d6lxp" Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.397936 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-d6lxp" event={"ID":"6fb3b18d-41af-427b-b2b1-e3b8cbbb0005","Type":"ContainerDied","Data":"8060fd6ef9343dadf328586c9faa2e712725728bf00bc6d3e58e77e0ae86007d"} Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.398026 4646 scope.go:117] "RemoveContainer" containerID="276ca750a898323d5ccda9ed35f43e91737a15239431d4b732e87c7f43128783" Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.401350 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ac255daf-f97d-47c1-9cfa-436577e24c50","Type":"ContainerStarted","Data":"3ae90c741ab29564cdc2e9ffdb4dfaa42a95e0e25327b46600d7dd3a11d5aca8"} Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.403181 4646 generic.go:334] "Generic (PLEG): container finished" podID="196f11c5-52de-42c3-ac45-dbaad2588b1d" containerID="516220b559295f4349fd056e4928f5d6622bdb69d1c8724ef0b206c458265b36" exitCode=0 Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.403245 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" event={"ID":"196f11c5-52de-42c3-ac45-dbaad2588b1d","Type":"ContainerDied","Data":"516220b559295f4349fd056e4928f5d6622bdb69d1c8724ef0b206c458265b36"} Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.403263 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" event={"ID":"196f11c5-52de-42c3-ac45-dbaad2588b1d","Type":"ContainerStarted","Data":"058092684cf790a447fd7368d9528a6d442431a79d2415a55c97334a3371408d"} Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.405170 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a559e2e6-4fbc-4c51-99fb-5e971142d41e","Type":"ContainerStarted","Data":"487f9b5d94f2404bbca651f0c2cec97ba35b47e2bfa264b0d4278db7e96f9e9a"} Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.465370 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.092013955 podStartE2EDuration="6.46534451s" podCreationTimestamp="2025-12-03 11:13:25 +0000 UTC" firstStartedPulling="2025-12-03 11:13:26.561986174 +0000 UTC m=+1183.025042309" lastFinishedPulling="2025-12-03 11:13:29.935316729 +0000 UTC m=+1186.398372864" observedRunningTime="2025-12-03 11:13:31.423937299 +0000 UTC m=+1187.886993434" watchObservedRunningTime="2025-12-03 11:13:31.46534451 +0000 UTC m=+1187.928400645" Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.492766 4646 scope.go:117] "RemoveContainer" containerID="bde7d201205a24673a3153a1ecd13230176ae3bfac1bfe5cde58e9241d4f2437" Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.494669 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-d6lxp"] Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.501533 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-d6lxp"] Dec 03 11:13:31 crc kubenswrapper[4646]: I1203 11:13:31.893598 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fb3b18d-41af-427b-b2b1-e3b8cbbb0005" path="/var/lib/kubelet/pods/6fb3b18d-41af-427b-b2b1-e3b8cbbb0005/volumes" Dec 03 11:13:32 crc kubenswrapper[4646]: I1203 11:13:32.448232 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" event={"ID":"196f11c5-52de-42c3-ac45-dbaad2588b1d","Type":"ContainerStarted","Data":"d1bde7db1f0b6ab39d82c1db108e699eb7edf462ede886515bb85cd726915308"} Dec 03 11:13:32 crc kubenswrapper[4646]: I1203 11:13:32.448584 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:32 crc kubenswrapper[4646]: I1203 11:13:32.457318 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a559e2e6-4fbc-4c51-99fb-5e971142d41e","Type":"ContainerStarted","Data":"d1dcde71ba498ce91f9efa7ca9dfefc43ef6daada1b6c97a966650feb4e99fd8"} Dec 03 11:13:32 crc kubenswrapper[4646]: I1203 11:13:32.480638 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" podStartSLOduration=3.480619019 podStartE2EDuration="3.480619019s" podCreationTimestamp="2025-12-03 11:13:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:13:32.47332059 +0000 UTC m=+1188.936376745" watchObservedRunningTime="2025-12-03 11:13:32.480619019 +0000 UTC m=+1188.943675144" Dec 03 11:13:32 crc kubenswrapper[4646]: I1203 11:13:32.780537 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 11:13:33 crc kubenswrapper[4646]: I1203 11:13:33.469433 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ac255daf-f97d-47c1-9cfa-436577e24c50","Type":"ContainerStarted","Data":"19fa29002d37e34da2abb6cfec6c7819f052ee2c39a0ade55d603bea5aab9ce5"} Dec 03 11:13:33 crc kubenswrapper[4646]: I1203 11:13:33.474729 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a559e2e6-4fbc-4c51-99fb-5e971142d41e","Type":"ContainerStarted","Data":"36bf63d44d74477babbf166dbefd3259be02a534d8b1c98399f5d279e99b6fcb"} Dec 03 11:13:33 crc kubenswrapper[4646]: I1203 11:13:33.474901 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a559e2e6-4fbc-4c51-99fb-5e971142d41e" containerName="cinder-api-log" containerID="cri-o://d1dcde71ba498ce91f9efa7ca9dfefc43ef6daada1b6c97a966650feb4e99fd8" gracePeriod=30 Dec 03 11:13:33 crc kubenswrapper[4646]: I1203 11:13:33.475319 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a559e2e6-4fbc-4c51-99fb-5e971142d41e" containerName="cinder-api" containerID="cri-o://36bf63d44d74477babbf166dbefd3259be02a534d8b1c98399f5d279e99b6fcb" gracePeriod=30 Dec 03 11:13:33 crc kubenswrapper[4646]: I1203 11:13:33.544835 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.544391794 podStartE2EDuration="4.544391794s" podCreationTimestamp="2025-12-03 11:13:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:13:33.491763997 +0000 UTC m=+1189.954820152" watchObservedRunningTime="2025-12-03 11:13:33.544391794 +0000 UTC m=+1190.007447939" Dec 03 11:13:34 crc kubenswrapper[4646]: I1203 11:13:34.336581 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:34 crc kubenswrapper[4646]: I1203 11:13:34.483163 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ac255daf-f97d-47c1-9cfa-436577e24c50","Type":"ContainerStarted","Data":"774b43b502516d953ef7ee0d70adca474df3cad73ee9941a88f383fd7ec5c08b"} Dec 03 11:13:34 crc kubenswrapper[4646]: I1203 11:13:34.485806 4646 generic.go:334] "Generic (PLEG): container finished" podID="a559e2e6-4fbc-4c51-99fb-5e971142d41e" containerID="d1dcde71ba498ce91f9efa7ca9dfefc43ef6daada1b6c97a966650feb4e99fd8" exitCode=143 Dec 03 11:13:34 crc kubenswrapper[4646]: I1203 11:13:34.485875 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a559e2e6-4fbc-4c51-99fb-5e971142d41e","Type":"ContainerDied","Data":"d1dcde71ba498ce91f9efa7ca9dfefc43ef6daada1b6c97a966650feb4e99fd8"} Dec 03 11:13:34 crc kubenswrapper[4646]: I1203 11:13:34.947403 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 11:13:35 crc kubenswrapper[4646]: I1203 11:13:35.059844 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:35 crc kubenswrapper[4646]: I1203 11:13:35.077111 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.070669458 podStartE2EDuration="6.077094967s" podCreationTimestamp="2025-12-03 11:13:29 +0000 UTC" firstStartedPulling="2025-12-03 11:13:30.943764053 +0000 UTC m=+1187.406820188" lastFinishedPulling="2025-12-03 11:13:31.950189562 +0000 UTC m=+1188.413245697" observedRunningTime="2025-12-03 11:13:34.50898359 +0000 UTC m=+1190.972039725" watchObservedRunningTime="2025-12-03 11:13:35.077094967 +0000 UTC m=+1191.540151102" Dec 03 11:13:35 crc kubenswrapper[4646]: I1203 11:13:35.409011 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 11:13:37 crc kubenswrapper[4646]: I1203 11:13:37.648140 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:37 crc kubenswrapper[4646]: I1203 11:13:37.697783 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6d775b8f86-64hmf" Dec 03 11:13:37 crc kubenswrapper[4646]: I1203 11:13:37.755265 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-c8d5f4d66-7z5sv"] Dec 03 11:13:37 crc kubenswrapper[4646]: I1203 11:13:37.755564 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-c8d5f4d66-7z5sv" podUID="b3d95cd5-dac1-4632-bb55-74af55174c0b" containerName="barbican-api-log" containerID="cri-o://b43f7e2e0255482a5fd87e63eee44cd517625e027c736cace5068dea0816d86b" gracePeriod=30 Dec 03 11:13:37 crc kubenswrapper[4646]: I1203 11:13:37.756375 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-c8d5f4d66-7z5sv" podUID="b3d95cd5-dac1-4632-bb55-74af55174c0b" containerName="barbican-api" containerID="cri-o://043dab58c03b248330030af1fe35f42b1c39d73c3d673b184e94d224adb8347c" gracePeriod=30 Dec 03 11:13:37 crc kubenswrapper[4646]: I1203 11:13:37.767721 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-c8d5f4d66-7z5sv" podUID="b3d95cd5-dac1-4632-bb55-74af55174c0b" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.145:9311/healthcheck\": EOF" Dec 03 11:13:38 crc kubenswrapper[4646]: I1203 11:13:38.523176 4646 generic.go:334] "Generic (PLEG): container finished" podID="b3d95cd5-dac1-4632-bb55-74af55174c0b" containerID="b43f7e2e0255482a5fd87e63eee44cd517625e027c736cace5068dea0816d86b" exitCode=143 Dec 03 11:13:38 crc kubenswrapper[4646]: I1203 11:13:38.523616 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c8d5f4d66-7z5sv" event={"ID":"b3d95cd5-dac1-4632-bb55-74af55174c0b","Type":"ContainerDied","Data":"b43f7e2e0255482a5fd87e63eee44cd517625e027c736cace5068dea0816d86b"} Dec 03 11:13:39 crc kubenswrapper[4646]: I1203 11:13:39.060542 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:39 crc kubenswrapper[4646]: I1203 11:13:39.077880 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:40 crc kubenswrapper[4646]: I1203 11:13:40.159282 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 11:13:40 crc kubenswrapper[4646]: I1203 11:13:40.207927 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 11:13:40 crc kubenswrapper[4646]: I1203 11:13:40.310571 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:40 crc kubenswrapper[4646]: I1203 11:13:40.311530 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-96574894d-gjvg2" Dec 03 11:13:40 crc kubenswrapper[4646]: I1203 11:13:40.357491 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:13:40 crc kubenswrapper[4646]: I1203 11:13:40.445923 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf"] Dec 03 11:13:40 crc kubenswrapper[4646]: I1203 11:13:40.446206 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" podUID="1495a64c-4fe7-4875-ae31-05ac0f3949b8" containerName="dnsmasq-dns" containerID="cri-o://91a79db0696da530f2cc08a85addb1b08293ef5ce79f580b8814f87f063caa45" gracePeriod=10 Dec 03 11:13:40 crc kubenswrapper[4646]: I1203 11:13:40.588247 4646 generic.go:334] "Generic (PLEG): container finished" podID="1495a64c-4fe7-4875-ae31-05ac0f3949b8" containerID="91a79db0696da530f2cc08a85addb1b08293ef5ce79f580b8814f87f063caa45" exitCode=0 Dec 03 11:13:40 crc kubenswrapper[4646]: I1203 11:13:40.588961 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="ac255daf-f97d-47c1-9cfa-436577e24c50" containerName="cinder-scheduler" containerID="cri-o://19fa29002d37e34da2abb6cfec6c7819f052ee2c39a0ade55d603bea5aab9ce5" gracePeriod=30 Dec 03 11:13:40 crc kubenswrapper[4646]: I1203 11:13:40.589447 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" event={"ID":"1495a64c-4fe7-4875-ae31-05ac0f3949b8","Type":"ContainerDied","Data":"91a79db0696da530f2cc08a85addb1b08293ef5ce79f580b8814f87f063caa45"} Dec 03 11:13:40 crc kubenswrapper[4646]: I1203 11:13:40.603035 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="ac255daf-f97d-47c1-9cfa-436577e24c50" containerName="probe" containerID="cri-o://774b43b502516d953ef7ee0d70adca474df3cad73ee9941a88f383fd7ec5c08b" gracePeriod=30 Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.307043 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.473958 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-config\") pod \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.474035 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-ovsdbserver-sb\") pod \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.474081 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-ovsdbserver-nb\") pod \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.474121 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5m7w\" (UniqueName: \"kubernetes.io/projected/1495a64c-4fe7-4875-ae31-05ac0f3949b8-kube-api-access-h5m7w\") pod \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.474260 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-dns-svc\") pod \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\" (UID: \"1495a64c-4fe7-4875-ae31-05ac0f3949b8\") " Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.503041 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1495a64c-4fe7-4875-ae31-05ac0f3949b8-kube-api-access-h5m7w" (OuterVolumeSpecName: "kube-api-access-h5m7w") pod "1495a64c-4fe7-4875-ae31-05ac0f3949b8" (UID: "1495a64c-4fe7-4875-ae31-05ac0f3949b8"). InnerVolumeSpecName "kube-api-access-h5m7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.541167 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-config" (OuterVolumeSpecName: "config") pod "1495a64c-4fe7-4875-ae31-05ac0f3949b8" (UID: "1495a64c-4fe7-4875-ae31-05ac0f3949b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.543493 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1495a64c-4fe7-4875-ae31-05ac0f3949b8" (UID: "1495a64c-4fe7-4875-ae31-05ac0f3949b8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.558024 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1495a64c-4fe7-4875-ae31-05ac0f3949b8" (UID: "1495a64c-4fe7-4875-ae31-05ac0f3949b8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:41 crc kubenswrapper[4646]: E1203 11:13:41.561871 4646 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac255daf_f97d_47c1_9cfa_436577e24c50.slice/crio-774b43b502516d953ef7ee0d70adca474df3cad73ee9941a88f383fd7ec5c08b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac255daf_f97d_47c1_9cfa_436577e24c50.slice/crio-conmon-774b43b502516d953ef7ee0d70adca474df3cad73ee9941a88f383fd7ec5c08b.scope\": RecentStats: unable to find data in memory cache]" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.578398 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.578426 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5m7w\" (UniqueName: \"kubernetes.io/projected/1495a64c-4fe7-4875-ae31-05ac0f3949b8-kube-api-access-h5m7w\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.578438 4646 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.578448 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.589983 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1495a64c-4fe7-4875-ae31-05ac0f3949b8" (UID: "1495a64c-4fe7-4875-ae31-05ac0f3949b8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.601325 4646 generic.go:334] "Generic (PLEG): container finished" podID="ac255daf-f97d-47c1-9cfa-436577e24c50" containerID="774b43b502516d953ef7ee0d70adca474df3cad73ee9941a88f383fd7ec5c08b" exitCode=0 Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.601378 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ac255daf-f97d-47c1-9cfa-436577e24c50","Type":"ContainerDied","Data":"774b43b502516d953ef7ee0d70adca474df3cad73ee9941a88f383fd7ec5c08b"} Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.605054 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" event={"ID":"1495a64c-4fe7-4875-ae31-05ac0f3949b8","Type":"ContainerDied","Data":"6cd550227fb32b9ba35b80943e871c1e7bc303b1dc2a4666e6f555f80ae49dd0"} Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.605124 4646 scope.go:117] "RemoveContainer" containerID="91a79db0696da530f2cc08a85addb1b08293ef5ce79f580b8814f87f063caa45" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.605169 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.618865 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-779dc4bcb9-gdhmd" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.662979 4646 scope.go:117] "RemoveContainer" containerID="a28909ef1d585195d0d1e36582d8c3ebf0c1cc15e627c2fefaae4e77a6f6c647" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.681968 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1495a64c-4fe7-4875-ae31-05ac0f3949b8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.683806 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf"] Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.712352 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-5wrhf"] Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.859744 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1495a64c-4fe7-4875-ae31-05ac0f3949b8" path="/var/lib/kubelet/pods/1495a64c-4fe7-4875-ae31-05ac0f3949b8/volumes" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.861316 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-59d75d495c-x9phz" Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.925954 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-9f5b87b8-m9k27"] Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.926234 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-9f5b87b8-m9k27" podUID="d0d65d94-d172-41e9-9e61-d44dd40dcb3e" containerName="neutron-api" containerID="cri-o://4b64af0103966c64399a94d2506ee2c8774c0cfe2acf8861a3d71d3ec4acd87e" gracePeriod=30 Dec 03 11:13:41 crc kubenswrapper[4646]: I1203 11:13:41.926310 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-9f5b87b8-m9k27" podUID="d0d65d94-d172-41e9-9e61-d44dd40dcb3e" containerName="neutron-httpd" containerID="cri-o://4099cdb499116765b3041174066958a5a667f51efa1143d60d22231cb86a282e" gracePeriod=30 Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.200275 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-c8d5f4d66-7z5sv" podUID="b3d95cd5-dac1-4632-bb55-74af55174c0b" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.145:9311/healthcheck\": read tcp 10.217.0.2:55516->10.217.0.145:9311: read: connection reset by peer" Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.200603 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-c8d5f4d66-7z5sv" podUID="b3d95cd5-dac1-4632-bb55-74af55174c0b" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.145:9311/healthcheck\": read tcp 10.217.0.2:55518->10.217.0.145:9311: read: connection reset by peer" Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.620972 4646 generic.go:334] "Generic (PLEG): container finished" podID="b3d95cd5-dac1-4632-bb55-74af55174c0b" containerID="043dab58c03b248330030af1fe35f42b1c39d73c3d673b184e94d224adb8347c" exitCode=0 Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.621185 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c8d5f4d66-7z5sv" event={"ID":"b3d95cd5-dac1-4632-bb55-74af55174c0b","Type":"ContainerDied","Data":"043dab58c03b248330030af1fe35f42b1c39d73c3d673b184e94d224adb8347c"} Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.635843 4646 generic.go:334] "Generic (PLEG): container finished" podID="d0d65d94-d172-41e9-9e61-d44dd40dcb3e" containerID="4099cdb499116765b3041174066958a5a667f51efa1143d60d22231cb86a282e" exitCode=0 Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.635901 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9f5b87b8-m9k27" event={"ID":"d0d65d94-d172-41e9-9e61-d44dd40dcb3e","Type":"ContainerDied","Data":"4099cdb499116765b3041174066958a5a667f51efa1143d60d22231cb86a282e"} Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.712234 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.907291 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-config-data\") pod \"b3d95cd5-dac1-4632-bb55-74af55174c0b\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.907534 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srb85\" (UniqueName: \"kubernetes.io/projected/b3d95cd5-dac1-4632-bb55-74af55174c0b-kube-api-access-srb85\") pod \"b3d95cd5-dac1-4632-bb55-74af55174c0b\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.907717 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3d95cd5-dac1-4632-bb55-74af55174c0b-logs\") pod \"b3d95cd5-dac1-4632-bb55-74af55174c0b\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.907829 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-config-data-custom\") pod \"b3d95cd5-dac1-4632-bb55-74af55174c0b\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.907967 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-combined-ca-bundle\") pod \"b3d95cd5-dac1-4632-bb55-74af55174c0b\" (UID: \"b3d95cd5-dac1-4632-bb55-74af55174c0b\") " Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.908514 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3d95cd5-dac1-4632-bb55-74af55174c0b-logs" (OuterVolumeSpecName: "logs") pod "b3d95cd5-dac1-4632-bb55-74af55174c0b" (UID: "b3d95cd5-dac1-4632-bb55-74af55174c0b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.913073 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b3d95cd5-dac1-4632-bb55-74af55174c0b" (UID: "b3d95cd5-dac1-4632-bb55-74af55174c0b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.934062 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3d95cd5-dac1-4632-bb55-74af55174c0b-kube-api-access-srb85" (OuterVolumeSpecName: "kube-api-access-srb85") pod "b3d95cd5-dac1-4632-bb55-74af55174c0b" (UID: "b3d95cd5-dac1-4632-bb55-74af55174c0b"). InnerVolumeSpecName "kube-api-access-srb85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.937459 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3d95cd5-dac1-4632-bb55-74af55174c0b" (UID: "b3d95cd5-dac1-4632-bb55-74af55174c0b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:42 crc kubenswrapper[4646]: I1203 11:13:42.960838 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-config-data" (OuterVolumeSpecName: "config-data") pod "b3d95cd5-dac1-4632-bb55-74af55174c0b" (UID: "b3d95cd5-dac1-4632-bb55-74af55174c0b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:43 crc kubenswrapper[4646]: I1203 11:13:43.009112 4646 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:43 crc kubenswrapper[4646]: I1203 11:13:43.009310 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:43 crc kubenswrapper[4646]: I1203 11:13:43.009403 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srb85\" (UniqueName: \"kubernetes.io/projected/b3d95cd5-dac1-4632-bb55-74af55174c0b-kube-api-access-srb85\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:43 crc kubenswrapper[4646]: I1203 11:13:43.009461 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d95cd5-dac1-4632-bb55-74af55174c0b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:43 crc kubenswrapper[4646]: I1203 11:13:43.009539 4646 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3d95cd5-dac1-4632-bb55-74af55174c0b-logs\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:43 crc kubenswrapper[4646]: I1203 11:13:43.412815 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 11:13:43 crc kubenswrapper[4646]: I1203 11:13:43.660688 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c8d5f4d66-7z5sv" event={"ID":"b3d95cd5-dac1-4632-bb55-74af55174c0b","Type":"ContainerDied","Data":"2067f23bb6007d03d172f339facf799c74e7af51cb98f73b509d0186d9235432"} Dec 03 11:13:43 crc kubenswrapper[4646]: I1203 11:13:43.660744 4646 scope.go:117] "RemoveContainer" containerID="043dab58c03b248330030af1fe35f42b1c39d73c3d673b184e94d224adb8347c" Dec 03 11:13:43 crc kubenswrapper[4646]: I1203 11:13:43.660762 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c8d5f4d66-7z5sv" Dec 03 11:13:43 crc kubenswrapper[4646]: I1203 11:13:43.704302 4646 scope.go:117] "RemoveContainer" containerID="b43f7e2e0255482a5fd87e63eee44cd517625e027c736cace5068dea0816d86b" Dec 03 11:13:43 crc kubenswrapper[4646]: I1203 11:13:43.713061 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-c8d5f4d66-7z5sv"] Dec 03 11:13:43 crc kubenswrapper[4646]: I1203 11:13:43.722676 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-c8d5f4d66-7z5sv"] Dec 03 11:13:43 crc kubenswrapper[4646]: I1203 11:13:43.872300 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3d95cd5-dac1-4632-bb55-74af55174c0b" path="/var/lib/kubelet/pods/b3d95cd5-dac1-4632-bb55-74af55174c0b/volumes" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.503293 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.548496 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-config-data-custom\") pod \"ac255daf-f97d-47c1-9cfa-436577e24c50\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.548565 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-config-data\") pod \"ac255daf-f97d-47c1-9cfa-436577e24c50\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.548592 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96k4l\" (UniqueName: \"kubernetes.io/projected/ac255daf-f97d-47c1-9cfa-436577e24c50-kube-api-access-96k4l\") pod \"ac255daf-f97d-47c1-9cfa-436577e24c50\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.548618 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-scripts\") pod \"ac255daf-f97d-47c1-9cfa-436577e24c50\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.549496 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ac255daf-f97d-47c1-9cfa-436577e24c50-etc-machine-id\") pod \"ac255daf-f97d-47c1-9cfa-436577e24c50\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.549518 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-combined-ca-bundle\") pod \"ac255daf-f97d-47c1-9cfa-436577e24c50\" (UID: \"ac255daf-f97d-47c1-9cfa-436577e24c50\") " Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.549752 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ac255daf-f97d-47c1-9cfa-436577e24c50-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ac255daf-f97d-47c1-9cfa-436577e24c50" (UID: "ac255daf-f97d-47c1-9cfa-436577e24c50"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.559781 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac255daf-f97d-47c1-9cfa-436577e24c50-kube-api-access-96k4l" (OuterVolumeSpecName: "kube-api-access-96k4l") pod "ac255daf-f97d-47c1-9cfa-436577e24c50" (UID: "ac255daf-f97d-47c1-9cfa-436577e24c50"). InnerVolumeSpecName "kube-api-access-96k4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.565704 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-scripts" (OuterVolumeSpecName: "scripts") pod "ac255daf-f97d-47c1-9cfa-436577e24c50" (UID: "ac255daf-f97d-47c1-9cfa-436577e24c50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.575449 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ac255daf-f97d-47c1-9cfa-436577e24c50" (UID: "ac255daf-f97d-47c1-9cfa-436577e24c50"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.639489 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac255daf-f97d-47c1-9cfa-436577e24c50" (UID: "ac255daf-f97d-47c1-9cfa-436577e24c50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.651389 4646 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ac255daf-f97d-47c1-9cfa-436577e24c50-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.651421 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.651430 4646 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.651442 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96k4l\" (UniqueName: \"kubernetes.io/projected/ac255daf-f97d-47c1-9cfa-436577e24c50-kube-api-access-96k4l\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.651451 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.663588 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-config-data" (OuterVolumeSpecName: "config-data") pod "ac255daf-f97d-47c1-9cfa-436577e24c50" (UID: "ac255daf-f97d-47c1-9cfa-436577e24c50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.690941 4646 generic.go:334] "Generic (PLEG): container finished" podID="ac255daf-f97d-47c1-9cfa-436577e24c50" containerID="19fa29002d37e34da2abb6cfec6c7819f052ee2c39a0ade55d603bea5aab9ce5" exitCode=0 Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.691136 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ac255daf-f97d-47c1-9cfa-436577e24c50","Type":"ContainerDied","Data":"19fa29002d37e34da2abb6cfec6c7819f052ee2c39a0ade55d603bea5aab9ce5"} Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.691209 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ac255daf-f97d-47c1-9cfa-436577e24c50","Type":"ContainerDied","Data":"3ae90c741ab29564cdc2e9ffdb4dfaa42a95e0e25327b46600d7dd3a11d5aca8"} Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.691277 4646 scope.go:117] "RemoveContainer" containerID="774b43b502516d953ef7ee0d70adca474df3cad73ee9941a88f383fd7ec5c08b" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.691445 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.723744 4646 scope.go:117] "RemoveContainer" containerID="19fa29002d37e34da2abb6cfec6c7819f052ee2c39a0ade55d603bea5aab9ce5" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.734964 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.753268 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac255daf-f97d-47c1-9cfa-436577e24c50-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.755389 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.772485 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 11:13:45 crc kubenswrapper[4646]: E1203 11:13:45.772929 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac255daf-f97d-47c1-9cfa-436577e24c50" containerName="probe" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.772946 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac255daf-f97d-47c1-9cfa-436577e24c50" containerName="probe" Dec 03 11:13:45 crc kubenswrapper[4646]: E1203 11:13:45.772959 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d95cd5-dac1-4632-bb55-74af55174c0b" containerName="barbican-api" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.772967 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d95cd5-dac1-4632-bb55-74af55174c0b" containerName="barbican-api" Dec 03 11:13:45 crc kubenswrapper[4646]: E1203 11:13:45.772977 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d95cd5-dac1-4632-bb55-74af55174c0b" containerName="barbican-api-log" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.772984 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d95cd5-dac1-4632-bb55-74af55174c0b" containerName="barbican-api-log" Dec 03 11:13:45 crc kubenswrapper[4646]: E1203 11:13:45.773003 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fb3b18d-41af-427b-b2b1-e3b8cbbb0005" containerName="dnsmasq-dns" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.773013 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fb3b18d-41af-427b-b2b1-e3b8cbbb0005" containerName="dnsmasq-dns" Dec 03 11:13:45 crc kubenswrapper[4646]: E1203 11:13:45.773031 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1495a64c-4fe7-4875-ae31-05ac0f3949b8" containerName="init" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.773039 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="1495a64c-4fe7-4875-ae31-05ac0f3949b8" containerName="init" Dec 03 11:13:45 crc kubenswrapper[4646]: E1203 11:13:45.773049 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1495a64c-4fe7-4875-ae31-05ac0f3949b8" containerName="dnsmasq-dns" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.773056 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="1495a64c-4fe7-4875-ae31-05ac0f3949b8" containerName="dnsmasq-dns" Dec 03 11:13:45 crc kubenswrapper[4646]: E1203 11:13:45.773073 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fb3b18d-41af-427b-b2b1-e3b8cbbb0005" containerName="init" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.773080 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fb3b18d-41af-427b-b2b1-e3b8cbbb0005" containerName="init" Dec 03 11:13:45 crc kubenswrapper[4646]: E1203 11:13:45.773097 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac255daf-f97d-47c1-9cfa-436577e24c50" containerName="cinder-scheduler" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.773106 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac255daf-f97d-47c1-9cfa-436577e24c50" containerName="cinder-scheduler" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.773312 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fb3b18d-41af-427b-b2b1-e3b8cbbb0005" containerName="dnsmasq-dns" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.773326 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="1495a64c-4fe7-4875-ae31-05ac0f3949b8" containerName="dnsmasq-dns" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.773364 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3d95cd5-dac1-4632-bb55-74af55174c0b" containerName="barbican-api" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.773378 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3d95cd5-dac1-4632-bb55-74af55174c0b" containerName="barbican-api-log" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.773390 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac255daf-f97d-47c1-9cfa-436577e24c50" containerName="cinder-scheduler" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.773407 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac255daf-f97d-47c1-9cfa-436577e24c50" containerName="probe" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.774502 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.782169 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.786533 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.794708 4646 scope.go:117] "RemoveContainer" containerID="774b43b502516d953ef7ee0d70adca474df3cad73ee9941a88f383fd7ec5c08b" Dec 03 11:13:45 crc kubenswrapper[4646]: E1203 11:13:45.802953 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"774b43b502516d953ef7ee0d70adca474df3cad73ee9941a88f383fd7ec5c08b\": container with ID starting with 774b43b502516d953ef7ee0d70adca474df3cad73ee9941a88f383fd7ec5c08b not found: ID does not exist" containerID="774b43b502516d953ef7ee0d70adca474df3cad73ee9941a88f383fd7ec5c08b" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.803147 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"774b43b502516d953ef7ee0d70adca474df3cad73ee9941a88f383fd7ec5c08b"} err="failed to get container status \"774b43b502516d953ef7ee0d70adca474df3cad73ee9941a88f383fd7ec5c08b\": rpc error: code = NotFound desc = could not find container \"774b43b502516d953ef7ee0d70adca474df3cad73ee9941a88f383fd7ec5c08b\": container with ID starting with 774b43b502516d953ef7ee0d70adca474df3cad73ee9941a88f383fd7ec5c08b not found: ID does not exist" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.803240 4646 scope.go:117] "RemoveContainer" containerID="19fa29002d37e34da2abb6cfec6c7819f052ee2c39a0ade55d603bea5aab9ce5" Dec 03 11:13:45 crc kubenswrapper[4646]: E1203 11:13:45.807240 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19fa29002d37e34da2abb6cfec6c7819f052ee2c39a0ade55d603bea5aab9ce5\": container with ID starting with 19fa29002d37e34da2abb6cfec6c7819f052ee2c39a0ade55d603bea5aab9ce5 not found: ID does not exist" containerID="19fa29002d37e34da2abb6cfec6c7819f052ee2c39a0ade55d603bea5aab9ce5" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.807273 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19fa29002d37e34da2abb6cfec6c7819f052ee2c39a0ade55d603bea5aab9ce5"} err="failed to get container status \"19fa29002d37e34da2abb6cfec6c7819f052ee2c39a0ade55d603bea5aab9ce5\": rpc error: code = NotFound desc = could not find container \"19fa29002d37e34da2abb6cfec6c7819f052ee2c39a0ade55d603bea5aab9ce5\": container with ID starting with 19fa29002d37e34da2abb6cfec6c7819f052ee2c39a0ade55d603bea5aab9ce5 not found: ID does not exist" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.857687 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e47516a7-a5ef-44f9-883e-b8c0383851c1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.857940 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e47516a7-a5ef-44f9-883e-b8c0383851c1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.858072 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e47516a7-a5ef-44f9-883e-b8c0383851c1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.858462 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e47516a7-a5ef-44f9-883e-b8c0383851c1-config-data\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.858557 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e47516a7-a5ef-44f9-883e-b8c0383851c1-scripts\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.863512 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j2hs\" (UniqueName: \"kubernetes.io/projected/e47516a7-a5ef-44f9-883e-b8c0383851c1-kube-api-access-4j2hs\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.869187 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac255daf-f97d-47c1-9cfa-436577e24c50" path="/var/lib/kubelet/pods/ac255daf-f97d-47c1-9cfa-436577e24c50/volumes" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.971247 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e47516a7-a5ef-44f9-883e-b8c0383851c1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.971688 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e47516a7-a5ef-44f9-883e-b8c0383851c1-config-data\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.971793 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e47516a7-a5ef-44f9-883e-b8c0383851c1-scripts\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.971930 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j2hs\" (UniqueName: \"kubernetes.io/projected/e47516a7-a5ef-44f9-883e-b8c0383851c1-kube-api-access-4j2hs\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.972153 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e47516a7-a5ef-44f9-883e-b8c0383851c1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.972230 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e47516a7-a5ef-44f9-883e-b8c0383851c1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.972236 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e47516a7-a5ef-44f9-883e-b8c0383851c1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.978863 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e47516a7-a5ef-44f9-883e-b8c0383851c1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.979400 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e47516a7-a5ef-44f9-883e-b8c0383851c1-config-data\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.988138 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e47516a7-a5ef-44f9-883e-b8c0383851c1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.988547 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e47516a7-a5ef-44f9-883e-b8c0383851c1-scripts\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:45 crc kubenswrapper[4646]: I1203 11:13:45.992014 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j2hs\" (UniqueName: \"kubernetes.io/projected/e47516a7-a5ef-44f9-883e-b8c0383851c1-kube-api-access-4j2hs\") pod \"cinder-scheduler-0\" (UID: \"e47516a7-a5ef-44f9-883e-b8c0383851c1\") " pod="openstack/cinder-scheduler-0" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.095677 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.574474 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.576216 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.579615 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.586271 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-qxw7s" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.586271 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.598885 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.615256 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.684221 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpbn7\" (UniqueName: \"kubernetes.io/projected/57b7374c-bde7-4db4-8dc3-897ff5ef6937-kube-api-access-wpbn7\") pod \"openstackclient\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.684313 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b7374c-bde7-4db4-8dc3-897ff5ef6937-combined-ca-bundle\") pod \"openstackclient\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.684378 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/57b7374c-bde7-4db4-8dc3-897ff5ef6937-openstack-config-secret\") pod \"openstackclient\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.684458 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/57b7374c-bde7-4db4-8dc3-897ff5ef6937-openstack-config\") pod \"openstackclient\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.705150 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e47516a7-a5ef-44f9-883e-b8c0383851c1","Type":"ContainerStarted","Data":"5426390c00ce93646e58ed3f3db929dd3ef2400c92ef6c97a9dfaf75fffc3e2f"} Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.786480 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpbn7\" (UniqueName: \"kubernetes.io/projected/57b7374c-bde7-4db4-8dc3-897ff5ef6937-kube-api-access-wpbn7\") pod \"openstackclient\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.786601 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b7374c-bde7-4db4-8dc3-897ff5ef6937-combined-ca-bundle\") pod \"openstackclient\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.787165 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/57b7374c-bde7-4db4-8dc3-897ff5ef6937-openstack-config-secret\") pod \"openstackclient\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.787255 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/57b7374c-bde7-4db4-8dc3-897ff5ef6937-openstack-config\") pod \"openstackclient\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.788203 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/57b7374c-bde7-4db4-8dc3-897ff5ef6937-openstack-config\") pod \"openstackclient\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.791646 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/57b7374c-bde7-4db4-8dc3-897ff5ef6937-openstack-config-secret\") pod \"openstackclient\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.793942 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b7374c-bde7-4db4-8dc3-897ff5ef6937-combined-ca-bundle\") pod \"openstackclient\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.805360 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpbn7\" (UniqueName: \"kubernetes.io/projected/57b7374c-bde7-4db4-8dc3-897ff5ef6937-kube-api-access-wpbn7\") pod \"openstackclient\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.890852 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.947531 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.952790 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.971122 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.978259 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.993283 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c18061a-8b6e-4c46-9475-2a2fbba4bd7a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4c18061a-8b6e-4c46-9475-2a2fbba4bd7a\") " pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.993352 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw4b4\" (UniqueName: \"kubernetes.io/projected/4c18061a-8b6e-4c46-9475-2a2fbba4bd7a-kube-api-access-mw4b4\") pod \"openstackclient\" (UID: \"4c18061a-8b6e-4c46-9475-2a2fbba4bd7a\") " pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.993509 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4c18061a-8b6e-4c46-9475-2a2fbba4bd7a-openstack-config\") pod \"openstackclient\" (UID: \"4c18061a-8b6e-4c46-9475-2a2fbba4bd7a\") " pod="openstack/openstackclient" Dec 03 11:13:46 crc kubenswrapper[4646]: I1203 11:13:46.993539 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4c18061a-8b6e-4c46-9475-2a2fbba4bd7a-openstack-config-secret\") pod \"openstackclient\" (UID: \"4c18061a-8b6e-4c46-9475-2a2fbba4bd7a\") " pod="openstack/openstackclient" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.007491 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 11:13:47 crc kubenswrapper[4646]: E1203 11:13:47.064613 4646 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 03 11:13:47 crc kubenswrapper[4646]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_57b7374c-bde7-4db4-8dc3-897ff5ef6937_0(0ece81701bd3d8a5ecd8ca720a770f43fe7bde88650b61705db4e838c25e7816): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"0ece81701bd3d8a5ecd8ca720a770f43fe7bde88650b61705db4e838c25e7816" Netns:"/var/run/netns/53a6a045-3e0d-46be-bba0-7f12376f05e8" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=0ece81701bd3d8a5ecd8ca720a770f43fe7bde88650b61705db4e838c25e7816;K8S_POD_UID=57b7374c-bde7-4db4-8dc3-897ff5ef6937" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/57b7374c-bde7-4db4-8dc3-897ff5ef6937]: expected pod UID "57b7374c-bde7-4db4-8dc3-897ff5ef6937" but got "4c18061a-8b6e-4c46-9475-2a2fbba4bd7a" from Kube API Dec 03 11:13:47 crc kubenswrapper[4646]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 11:13:47 crc kubenswrapper[4646]: > Dec 03 11:13:47 crc kubenswrapper[4646]: E1203 11:13:47.064674 4646 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 03 11:13:47 crc kubenswrapper[4646]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_57b7374c-bde7-4db4-8dc3-897ff5ef6937_0(0ece81701bd3d8a5ecd8ca720a770f43fe7bde88650b61705db4e838c25e7816): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"0ece81701bd3d8a5ecd8ca720a770f43fe7bde88650b61705db4e838c25e7816" Netns:"/var/run/netns/53a6a045-3e0d-46be-bba0-7f12376f05e8" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=0ece81701bd3d8a5ecd8ca720a770f43fe7bde88650b61705db4e838c25e7816;K8S_POD_UID=57b7374c-bde7-4db4-8dc3-897ff5ef6937" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/57b7374c-bde7-4db4-8dc3-897ff5ef6937]: expected pod UID "57b7374c-bde7-4db4-8dc3-897ff5ef6937" but got "4c18061a-8b6e-4c46-9475-2a2fbba4bd7a" from Kube API Dec 03 11:13:47 crc kubenswrapper[4646]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 11:13:47 crc kubenswrapper[4646]: > pod="openstack/openstackclient" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.096323 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw4b4\" (UniqueName: \"kubernetes.io/projected/4c18061a-8b6e-4c46-9475-2a2fbba4bd7a-kube-api-access-mw4b4\") pod \"openstackclient\" (UID: \"4c18061a-8b6e-4c46-9475-2a2fbba4bd7a\") " pod="openstack/openstackclient" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.096532 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4c18061a-8b6e-4c46-9475-2a2fbba4bd7a-openstack-config\") pod \"openstackclient\" (UID: \"4c18061a-8b6e-4c46-9475-2a2fbba4bd7a\") " pod="openstack/openstackclient" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.096559 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4c18061a-8b6e-4c46-9475-2a2fbba4bd7a-openstack-config-secret\") pod \"openstackclient\" (UID: \"4c18061a-8b6e-4c46-9475-2a2fbba4bd7a\") " pod="openstack/openstackclient" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.096684 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c18061a-8b6e-4c46-9475-2a2fbba4bd7a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4c18061a-8b6e-4c46-9475-2a2fbba4bd7a\") " pod="openstack/openstackclient" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.097901 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4c18061a-8b6e-4c46-9475-2a2fbba4bd7a-openstack-config\") pod \"openstackclient\" (UID: \"4c18061a-8b6e-4c46-9475-2a2fbba4bd7a\") " pod="openstack/openstackclient" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.107567 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c18061a-8b6e-4c46-9475-2a2fbba4bd7a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4c18061a-8b6e-4c46-9475-2a2fbba4bd7a\") " pod="openstack/openstackclient" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.118090 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4c18061a-8b6e-4c46-9475-2a2fbba4bd7a-openstack-config-secret\") pod \"openstackclient\" (UID: \"4c18061a-8b6e-4c46-9475-2a2fbba4bd7a\") " pod="openstack/openstackclient" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.121181 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw4b4\" (UniqueName: \"kubernetes.io/projected/4c18061a-8b6e-4c46-9475-2a2fbba4bd7a-kube-api-access-mw4b4\") pod \"openstackclient\" (UID: \"4c18061a-8b6e-4c46-9475-2a2fbba4bd7a\") " pod="openstack/openstackclient" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.305664 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.489044 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.505003 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-combined-ca-bundle\") pod \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.505139 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-ovndb-tls-certs\") pod \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.505175 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-config\") pod \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.505231 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5jk6\" (UniqueName: \"kubernetes.io/projected/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-kube-api-access-v5jk6\") pod \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.505425 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-httpd-config\") pod \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\" (UID: \"d0d65d94-d172-41e9-9e61-d44dd40dcb3e\") " Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.544608 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "d0d65d94-d172-41e9-9e61-d44dd40dcb3e" (UID: "d0d65d94-d172-41e9-9e61-d44dd40dcb3e"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.565154 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-kube-api-access-v5jk6" (OuterVolumeSpecName: "kube-api-access-v5jk6") pod "d0d65d94-d172-41e9-9e61-d44dd40dcb3e" (UID: "d0d65d94-d172-41e9-9e61-d44dd40dcb3e"). InnerVolumeSpecName "kube-api-access-v5jk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.607306 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5jk6\" (UniqueName: \"kubernetes.io/projected/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-kube-api-access-v5jk6\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.611396 4646 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.739329 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0d65d94-d172-41e9-9e61-d44dd40dcb3e" (UID: "d0d65d94-d172-41e9-9e61-d44dd40dcb3e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.757249 4646 generic.go:334] "Generic (PLEG): container finished" podID="d0d65d94-d172-41e9-9e61-d44dd40dcb3e" containerID="4b64af0103966c64399a94d2506ee2c8774c0cfe2acf8861a3d71d3ec4acd87e" exitCode=0 Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.757434 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9f5b87b8-m9k27" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.757640 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9f5b87b8-m9k27" event={"ID":"d0d65d94-d172-41e9-9e61-d44dd40dcb3e","Type":"ContainerDied","Data":"4b64af0103966c64399a94d2506ee2c8774c0cfe2acf8861a3d71d3ec4acd87e"} Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.757692 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9f5b87b8-m9k27" event={"ID":"d0d65d94-d172-41e9-9e61-d44dd40dcb3e","Type":"ContainerDied","Data":"0b8a20a4736f375e0bbe9e808ae10768ab044ed2193046e4e6183f86481c50af"} Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.757712 4646 scope.go:117] "RemoveContainer" containerID="4099cdb499116765b3041174066958a5a667f51efa1143d60d22231cb86a282e" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.763671 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.769871 4646 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="57b7374c-bde7-4db4-8dc3-897ff5ef6937" podUID="4c18061a-8b6e-4c46-9475-2a2fbba4bd7a" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.771706 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-config" (OuterVolumeSpecName: "config") pod "d0d65d94-d172-41e9-9e61-d44dd40dcb3e" (UID: "d0d65d94-d172-41e9-9e61-d44dd40dcb3e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.778791 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.793283 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "d0d65d94-d172-41e9-9e61-d44dd40dcb3e" (UID: "d0d65d94-d172-41e9-9e61-d44dd40dcb3e"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.796755 4646 scope.go:117] "RemoveContainer" containerID="4b64af0103966c64399a94d2506ee2c8774c0cfe2acf8861a3d71d3ec4acd87e" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.813682 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b7374c-bde7-4db4-8dc3-897ff5ef6937-combined-ca-bundle\") pod \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.813835 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpbn7\" (UniqueName: \"kubernetes.io/projected/57b7374c-bde7-4db4-8dc3-897ff5ef6937-kube-api-access-wpbn7\") pod \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.813887 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/57b7374c-bde7-4db4-8dc3-897ff5ef6937-openstack-config-secret\") pod \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.813926 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/57b7374c-bde7-4db4-8dc3-897ff5ef6937-openstack-config\") pod \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\" (UID: \"57b7374c-bde7-4db4-8dc3-897ff5ef6937\") " Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.814211 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.814223 4646 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.814232 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d0d65d94-d172-41e9-9e61-d44dd40dcb3e-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.816227 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57b7374c-bde7-4db4-8dc3-897ff5ef6937-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "57b7374c-bde7-4db4-8dc3-897ff5ef6937" (UID: "57b7374c-bde7-4db4-8dc3-897ff5ef6937"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.817282 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57b7374c-bde7-4db4-8dc3-897ff5ef6937-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57b7374c-bde7-4db4-8dc3-897ff5ef6937" (UID: "57b7374c-bde7-4db4-8dc3-897ff5ef6937"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.821564 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57b7374c-bde7-4db4-8dc3-897ff5ef6937-kube-api-access-wpbn7" (OuterVolumeSpecName: "kube-api-access-wpbn7") pod "57b7374c-bde7-4db4-8dc3-897ff5ef6937" (UID: "57b7374c-bde7-4db4-8dc3-897ff5ef6937"). InnerVolumeSpecName "kube-api-access-wpbn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.825525 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57b7374c-bde7-4db4-8dc3-897ff5ef6937-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "57b7374c-bde7-4db4-8dc3-897ff5ef6937" (UID: "57b7374c-bde7-4db4-8dc3-897ff5ef6937"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.825907 4646 scope.go:117] "RemoveContainer" containerID="4099cdb499116765b3041174066958a5a667f51efa1143d60d22231cb86a282e" Dec 03 11:13:47 crc kubenswrapper[4646]: E1203 11:13:47.828022 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4099cdb499116765b3041174066958a5a667f51efa1143d60d22231cb86a282e\": container with ID starting with 4099cdb499116765b3041174066958a5a667f51efa1143d60d22231cb86a282e not found: ID does not exist" containerID="4099cdb499116765b3041174066958a5a667f51efa1143d60d22231cb86a282e" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.828078 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4099cdb499116765b3041174066958a5a667f51efa1143d60d22231cb86a282e"} err="failed to get container status \"4099cdb499116765b3041174066958a5a667f51efa1143d60d22231cb86a282e\": rpc error: code = NotFound desc = could not find container \"4099cdb499116765b3041174066958a5a667f51efa1143d60d22231cb86a282e\": container with ID starting with 4099cdb499116765b3041174066958a5a667f51efa1143d60d22231cb86a282e not found: ID does not exist" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.828112 4646 scope.go:117] "RemoveContainer" containerID="4b64af0103966c64399a94d2506ee2c8774c0cfe2acf8861a3d71d3ec4acd87e" Dec 03 11:13:47 crc kubenswrapper[4646]: E1203 11:13:47.828556 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b64af0103966c64399a94d2506ee2c8774c0cfe2acf8861a3d71d3ec4acd87e\": container with ID starting with 4b64af0103966c64399a94d2506ee2c8774c0cfe2acf8861a3d71d3ec4acd87e not found: ID does not exist" containerID="4b64af0103966c64399a94d2506ee2c8774c0cfe2acf8861a3d71d3ec4acd87e" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.828596 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b64af0103966c64399a94d2506ee2c8774c0cfe2acf8861a3d71d3ec4acd87e"} err="failed to get container status \"4b64af0103966c64399a94d2506ee2c8774c0cfe2acf8861a3d71d3ec4acd87e\": rpc error: code = NotFound desc = could not find container \"4b64af0103966c64399a94d2506ee2c8774c0cfe2acf8861a3d71d3ec4acd87e\": container with ID starting with 4b64af0103966c64399a94d2506ee2c8774c0cfe2acf8861a3d71d3ec4acd87e not found: ID does not exist" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.863870 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57b7374c-bde7-4db4-8dc3-897ff5ef6937" path="/var/lib/kubelet/pods/57b7374c-bde7-4db4-8dc3-897ff5ef6937/volumes" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.915719 4646 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/57b7374c-bde7-4db4-8dc3-897ff5ef6937-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.915762 4646 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/57b7374c-bde7-4db4-8dc3-897ff5ef6937-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.915772 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b7374c-bde7-4db4-8dc3-897ff5ef6937-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.915781 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpbn7\" (UniqueName: \"kubernetes.io/projected/57b7374c-bde7-4db4-8dc3-897ff5ef6937-kube-api-access-wpbn7\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:47 crc kubenswrapper[4646]: I1203 11:13:47.972648 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 11:13:48 crc kubenswrapper[4646]: I1203 11:13:48.099848 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-9f5b87b8-m9k27"] Dec 03 11:13:48 crc kubenswrapper[4646]: I1203 11:13:48.111522 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-9f5b87b8-m9k27"] Dec 03 11:13:48 crc kubenswrapper[4646]: I1203 11:13:48.776133 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"4c18061a-8b6e-4c46-9475-2a2fbba4bd7a","Type":"ContainerStarted","Data":"08911e758dfa18d7aff53ed83d42300fd8e5a660dba1837015a68636b66e7881"} Dec 03 11:13:48 crc kubenswrapper[4646]: I1203 11:13:48.796278 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 11:13:48 crc kubenswrapper[4646]: I1203 11:13:48.796279 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e47516a7-a5ef-44f9-883e-b8c0383851c1","Type":"ContainerStarted","Data":"2b2a35fca0ba2e71ba07e979ed599ac603482ac05e51d0e8e52b383b34c92e63"} Dec 03 11:13:48 crc kubenswrapper[4646]: I1203 11:13:48.796359 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e47516a7-a5ef-44f9-883e-b8c0383851c1","Type":"ContainerStarted","Data":"31576810f398164c1f37de599ec7e1118d3e0b84a833c357c3ef0ad0cfe803ad"} Dec 03 11:13:48 crc kubenswrapper[4646]: I1203 11:13:48.819310 4646 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="57b7374c-bde7-4db4-8dc3-897ff5ef6937" podUID="4c18061a-8b6e-4c46-9475-2a2fbba4bd7a" Dec 03 11:13:48 crc kubenswrapper[4646]: I1203 11:13:48.819907 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.819886764 podStartE2EDuration="3.819886764s" podCreationTimestamp="2025-12-03 11:13:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:13:48.816231704 +0000 UTC m=+1205.279287839" watchObservedRunningTime="2025-12-03 11:13:48.819886764 +0000 UTC m=+1205.282942899" Dec 03 11:13:49 crc kubenswrapper[4646]: I1203 11:13:49.868244 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0d65d94-d172-41e9-9e61-d44dd40dcb3e" path="/var/lib/kubelet/pods/d0d65d94-d172-41e9-9e61-d44dd40dcb3e/volumes" Dec 03 11:13:51 crc kubenswrapper[4646]: I1203 11:13:51.096747 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 11:13:55 crc kubenswrapper[4646]: I1203 11:13:55.964711 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:13:55 crc kubenswrapper[4646]: I1203 11:13:55.965319 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:13:56 crc kubenswrapper[4646]: I1203 11:13:56.074045 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 11:13:56 crc kubenswrapper[4646]: I1203 11:13:56.437694 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 11:13:56 crc kubenswrapper[4646]: I1203 11:13:56.659501 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:13:56 crc kubenswrapper[4646]: I1203 11:13:56.887260 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="ceilometer-central-agent" containerID="cri-o://e6968eaa9205bf5248c420c091e74ce4d23ab8f237df5357e8c27c18ea9d4f63" gracePeriod=30 Dec 03 11:13:56 crc kubenswrapper[4646]: I1203 11:13:56.887914 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="proxy-httpd" containerID="cri-o://d9f3e82110bbcce0aa356967fc9c8d8e8a9cae58f1a09379a288ecd647359094" gracePeriod=30 Dec 03 11:13:56 crc kubenswrapper[4646]: I1203 11:13:56.887970 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="sg-core" containerID="cri-o://10b5c6cd346c8e48d288dc5e747ab8d805b55ee8e20a99e32d248e1b17f53bac" gracePeriod=30 Dec 03 11:13:56 crc kubenswrapper[4646]: I1203 11:13:56.888003 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="ceilometer-notification-agent" containerID="cri-o://2106226de82078f155cc8466d5d9925b0cde38c2ebe6413d587efe6acff2c697" gracePeriod=30 Dec 03 11:13:57 crc kubenswrapper[4646]: I1203 11:13:57.903167 4646 generic.go:334] "Generic (PLEG): container finished" podID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerID="d9f3e82110bbcce0aa356967fc9c8d8e8a9cae58f1a09379a288ecd647359094" exitCode=0 Dec 03 11:13:57 crc kubenswrapper[4646]: I1203 11:13:57.903653 4646 generic.go:334] "Generic (PLEG): container finished" podID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerID="10b5c6cd346c8e48d288dc5e747ab8d805b55ee8e20a99e32d248e1b17f53bac" exitCode=2 Dec 03 11:13:57 crc kubenswrapper[4646]: I1203 11:13:57.903663 4646 generic.go:334] "Generic (PLEG): container finished" podID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerID="2106226de82078f155cc8466d5d9925b0cde38c2ebe6413d587efe6acff2c697" exitCode=0 Dec 03 11:13:57 crc kubenswrapper[4646]: I1203 11:13:57.903670 4646 generic.go:334] "Generic (PLEG): container finished" podID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerID="e6968eaa9205bf5248c420c091e74ce4d23ab8f237df5357e8c27c18ea9d4f63" exitCode=0 Dec 03 11:13:57 crc kubenswrapper[4646]: I1203 11:13:57.903231 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6c9f14af-75f4-4b2a-97fc-97aa12438933","Type":"ContainerDied","Data":"d9f3e82110bbcce0aa356967fc9c8d8e8a9cae58f1a09379a288ecd647359094"} Dec 03 11:13:57 crc kubenswrapper[4646]: I1203 11:13:57.903696 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6c9f14af-75f4-4b2a-97fc-97aa12438933","Type":"ContainerDied","Data":"10b5c6cd346c8e48d288dc5e747ab8d805b55ee8e20a99e32d248e1b17f53bac"} Dec 03 11:13:57 crc kubenswrapper[4646]: I1203 11:13:57.903705 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6c9f14af-75f4-4b2a-97fc-97aa12438933","Type":"ContainerDied","Data":"2106226de82078f155cc8466d5d9925b0cde38c2ebe6413d587efe6acff2c697"} Dec 03 11:13:57 crc kubenswrapper[4646]: I1203 11:13:57.903717 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6c9f14af-75f4-4b2a-97fc-97aa12438933","Type":"ContainerDied","Data":"e6968eaa9205bf5248c420c091e74ce4d23ab8f237df5357e8c27c18ea9d4f63"} Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.806889 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.865933 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-scripts\") pod \"6c9f14af-75f4-4b2a-97fc-97aa12438933\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.866002 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-config-data\") pod \"6c9f14af-75f4-4b2a-97fc-97aa12438933\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.866139 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c9f14af-75f4-4b2a-97fc-97aa12438933-run-httpd\") pod \"6c9f14af-75f4-4b2a-97fc-97aa12438933\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.866245 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-combined-ca-bundle\") pod \"6c9f14af-75f4-4b2a-97fc-97aa12438933\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.866287 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46btx\" (UniqueName: \"kubernetes.io/projected/6c9f14af-75f4-4b2a-97fc-97aa12438933-kube-api-access-46btx\") pod \"6c9f14af-75f4-4b2a-97fc-97aa12438933\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.866383 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-sg-core-conf-yaml\") pod \"6c9f14af-75f4-4b2a-97fc-97aa12438933\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.866415 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c9f14af-75f4-4b2a-97fc-97aa12438933-log-httpd\") pod \"6c9f14af-75f4-4b2a-97fc-97aa12438933\" (UID: \"6c9f14af-75f4-4b2a-97fc-97aa12438933\") " Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.868476 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c9f14af-75f4-4b2a-97fc-97aa12438933-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6c9f14af-75f4-4b2a-97fc-97aa12438933" (UID: "6c9f14af-75f4-4b2a-97fc-97aa12438933"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.870812 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c9f14af-75f4-4b2a-97fc-97aa12438933-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6c9f14af-75f4-4b2a-97fc-97aa12438933" (UID: "6c9f14af-75f4-4b2a-97fc-97aa12438933"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.903846 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c9f14af-75f4-4b2a-97fc-97aa12438933-kube-api-access-46btx" (OuterVolumeSpecName: "kube-api-access-46btx") pod "6c9f14af-75f4-4b2a-97fc-97aa12438933" (UID: "6c9f14af-75f4-4b2a-97fc-97aa12438933"). InnerVolumeSpecName "kube-api-access-46btx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.912409 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-scripts" (OuterVolumeSpecName: "scripts") pod "6c9f14af-75f4-4b2a-97fc-97aa12438933" (UID: "6c9f14af-75f4-4b2a-97fc-97aa12438933"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.913176 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"4c18061a-8b6e-4c46-9475-2a2fbba4bd7a","Type":"ContainerStarted","Data":"c5cece70c133d38b4c9e158463d766d1c698f4c3f3e16969b968c015fa875d1b"} Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.915392 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6c9f14af-75f4-4b2a-97fc-97aa12438933","Type":"ContainerDied","Data":"248c45089ab456646945e0c6a14adceffa4cd008bd2d5dcc864add7fb295cb5f"} Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.915448 4646 scope.go:117] "RemoveContainer" containerID="d9f3e82110bbcce0aa356967fc9c8d8e8a9cae58f1a09379a288ecd647359094" Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.915469 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.939611 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6c9f14af-75f4-4b2a-97fc-97aa12438933" (UID: "6c9f14af-75f4-4b2a-97fc-97aa12438933"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.975040 4646 scope.go:117] "RemoveContainer" containerID="10b5c6cd346c8e48d288dc5e747ab8d805b55ee8e20a99e32d248e1b17f53bac" Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.977280 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46btx\" (UniqueName: \"kubernetes.io/projected/6c9f14af-75f4-4b2a-97fc-97aa12438933-kube-api-access-46btx\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.977311 4646 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.977319 4646 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c9f14af-75f4-4b2a-97fc-97aa12438933-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.977328 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:58 crc kubenswrapper[4646]: I1203 11:13:58.977340 4646 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6c9f14af-75f4-4b2a-97fc-97aa12438933-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.013670 4646 scope.go:117] "RemoveContainer" containerID="2106226de82078f155cc8466d5d9925b0cde38c2ebe6413d587efe6acff2c697" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.052289 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-config-data" (OuterVolumeSpecName: "config-data") pod "6c9f14af-75f4-4b2a-97fc-97aa12438933" (UID: "6c9f14af-75f4-4b2a-97fc-97aa12438933"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.058671 4646 scope.go:117] "RemoveContainer" containerID="e6968eaa9205bf5248c420c091e74ce4d23ab8f237df5357e8c27c18ea9d4f63" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.088455 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.090380 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c9f14af-75f4-4b2a-97fc-97aa12438933" (UID: "6c9f14af-75f4-4b2a-97fc-97aa12438933"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.189924 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c9f14af-75f4-4b2a-97fc-97aa12438933-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.246599 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.7537217480000002 podStartE2EDuration="13.24657742s" podCreationTimestamp="2025-12-03 11:13:46 +0000 UTC" firstStartedPulling="2025-12-03 11:13:47.978229196 +0000 UTC m=+1204.441285341" lastFinishedPulling="2025-12-03 11:13:58.471084878 +0000 UTC m=+1214.934141013" observedRunningTime="2025-12-03 11:13:58.975319031 +0000 UTC m=+1215.438375156" watchObservedRunningTime="2025-12-03 11:13:59.24657742 +0000 UTC m=+1215.709633555" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.248048 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.258495 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.280519 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:13:59 crc kubenswrapper[4646]: E1203 11:13:59.280858 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="ceilometer-central-agent" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.280875 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="ceilometer-central-agent" Dec 03 11:13:59 crc kubenswrapper[4646]: E1203 11:13:59.280895 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d65d94-d172-41e9-9e61-d44dd40dcb3e" containerName="neutron-httpd" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.280902 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d65d94-d172-41e9-9e61-d44dd40dcb3e" containerName="neutron-httpd" Dec 03 11:13:59 crc kubenswrapper[4646]: E1203 11:13:59.280922 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="sg-core" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.280928 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="sg-core" Dec 03 11:13:59 crc kubenswrapper[4646]: E1203 11:13:59.280942 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="proxy-httpd" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.280948 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="proxy-httpd" Dec 03 11:13:59 crc kubenswrapper[4646]: E1203 11:13:59.280962 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d65d94-d172-41e9-9e61-d44dd40dcb3e" containerName="neutron-api" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.280968 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d65d94-d172-41e9-9e61-d44dd40dcb3e" containerName="neutron-api" Dec 03 11:13:59 crc kubenswrapper[4646]: E1203 11:13:59.280984 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="ceilometer-notification-agent" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.280990 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="ceilometer-notification-agent" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.281149 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="sg-core" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.281159 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="ceilometer-central-agent" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.281174 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0d65d94-d172-41e9-9e61-d44dd40dcb3e" containerName="neutron-api" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.281184 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0d65d94-d172-41e9-9e61-d44dd40dcb3e" containerName="neutron-httpd" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.281191 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="ceilometer-notification-agent" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.281198 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" containerName="proxy-httpd" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.283480 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.302120 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.302622 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.303987 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.393152 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/222c4fa4-6552-4021-ab4d-58f1a343a97d-log-httpd\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.393218 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-scripts\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.393269 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sztzw\" (UniqueName: \"kubernetes.io/projected/222c4fa4-6552-4021-ab4d-58f1a343a97d-kube-api-access-sztzw\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.393285 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/222c4fa4-6552-4021-ab4d-58f1a343a97d-run-httpd\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.393305 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-config-data\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.393370 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.393398 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.494731 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.494795 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.494862 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/222c4fa4-6552-4021-ab4d-58f1a343a97d-log-httpd\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.494925 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-scripts\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.495469 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/222c4fa4-6552-4021-ab4d-58f1a343a97d-log-httpd\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.495682 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sztzw\" (UniqueName: \"kubernetes.io/projected/222c4fa4-6552-4021-ab4d-58f1a343a97d-kube-api-access-sztzw\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.496070 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/222c4fa4-6552-4021-ab4d-58f1a343a97d-run-httpd\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.496096 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-config-data\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.496397 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/222c4fa4-6552-4021-ab4d-58f1a343a97d-run-httpd\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.498571 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-scripts\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.498943 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.503811 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.514201 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-config-data\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.515003 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sztzw\" (UniqueName: \"kubernetes.io/projected/222c4fa4-6552-4021-ab4d-58f1a343a97d-kube-api-access-sztzw\") pod \"ceilometer-0\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.543881 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.545205 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:13:59 crc kubenswrapper[4646]: I1203 11:13:59.859368 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c9f14af-75f4-4b2a-97fc-97aa12438933" path="/var/lib/kubelet/pods/6c9f14af-75f4-4b2a-97fc-97aa12438933/volumes" Dec 03 11:14:00 crc kubenswrapper[4646]: I1203 11:14:00.028223 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:14:00 crc kubenswrapper[4646]: I1203 11:14:00.930704 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"222c4fa4-6552-4021-ab4d-58f1a343a97d","Type":"ContainerStarted","Data":"92a29e17bde562ccf50740b4353f58418d356746e4b412f777f901691a62a737"} Dec 03 11:14:00 crc kubenswrapper[4646]: I1203 11:14:00.930936 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"222c4fa4-6552-4021-ab4d-58f1a343a97d","Type":"ContainerStarted","Data":"dab4f562a5dae6a79960c8fef75aad4b44c8af54f272ea14878b5617db65164a"} Dec 03 11:14:01 crc kubenswrapper[4646]: I1203 11:14:01.949848 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"222c4fa4-6552-4021-ab4d-58f1a343a97d","Type":"ContainerStarted","Data":"e8f098272126786a2cdc6fa03c7b634611a0e63b9ae1517186f739a749575709"} Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.554622 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.600329 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a559e2e6-4fbc-4c51-99fb-5e971142d41e-etc-machine-id\") pod \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.600419 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a559e2e6-4fbc-4c51-99fb-5e971142d41e-logs\") pod \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.600476 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-config-data\") pod \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.600524 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-scripts\") pod \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.600599 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-config-data-custom\") pod \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.600626 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-combined-ca-bundle\") pod \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.600674 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bbqk\" (UniqueName: \"kubernetes.io/projected/a559e2e6-4fbc-4c51-99fb-5e971142d41e-kube-api-access-7bbqk\") pod \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\" (UID: \"a559e2e6-4fbc-4c51-99fb-5e971142d41e\") " Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.607909 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a559e2e6-4fbc-4c51-99fb-5e971142d41e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a559e2e6-4fbc-4c51-99fb-5e971142d41e" (UID: "a559e2e6-4fbc-4c51-99fb-5e971142d41e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.608300 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a559e2e6-4fbc-4c51-99fb-5e971142d41e-logs" (OuterVolumeSpecName: "logs") pod "a559e2e6-4fbc-4c51-99fb-5e971142d41e" (UID: "a559e2e6-4fbc-4c51-99fb-5e971142d41e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.615767 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a559e2e6-4fbc-4c51-99fb-5e971142d41e" (UID: "a559e2e6-4fbc-4c51-99fb-5e971142d41e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.616110 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a559e2e6-4fbc-4c51-99fb-5e971142d41e-kube-api-access-7bbqk" (OuterVolumeSpecName: "kube-api-access-7bbqk") pod "a559e2e6-4fbc-4c51-99fb-5e971142d41e" (UID: "a559e2e6-4fbc-4c51-99fb-5e971142d41e"). InnerVolumeSpecName "kube-api-access-7bbqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.616667 4646 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a559e2e6-4fbc-4c51-99fb-5e971142d41e-logs\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.616769 4646 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.616826 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bbqk\" (UniqueName: \"kubernetes.io/projected/a559e2e6-4fbc-4c51-99fb-5e971142d41e-kube-api-access-7bbqk\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.616878 4646 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a559e2e6-4fbc-4c51-99fb-5e971142d41e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.620434 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-scripts" (OuterVolumeSpecName: "scripts") pod "a559e2e6-4fbc-4c51-99fb-5e971142d41e" (UID: "a559e2e6-4fbc-4c51-99fb-5e971142d41e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.671424 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a559e2e6-4fbc-4c51-99fb-5e971142d41e" (UID: "a559e2e6-4fbc-4c51-99fb-5e971142d41e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.690649 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-config-data" (OuterVolumeSpecName: "config-data") pod "a559e2e6-4fbc-4c51-99fb-5e971142d41e" (UID: "a559e2e6-4fbc-4c51-99fb-5e971142d41e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.719506 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.719543 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.719556 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a559e2e6-4fbc-4c51-99fb-5e971142d41e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.977094 4646 generic.go:334] "Generic (PLEG): container finished" podID="a559e2e6-4fbc-4c51-99fb-5e971142d41e" containerID="36bf63d44d74477babbf166dbefd3259be02a534d8b1c98399f5d279e99b6fcb" exitCode=137 Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.977161 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a559e2e6-4fbc-4c51-99fb-5e971142d41e","Type":"ContainerDied","Data":"36bf63d44d74477babbf166dbefd3259be02a534d8b1c98399f5d279e99b6fcb"} Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.977189 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.977788 4646 scope.go:117] "RemoveContainer" containerID="36bf63d44d74477babbf166dbefd3259be02a534d8b1c98399f5d279e99b6fcb" Dec 03 11:14:04 crc kubenswrapper[4646]: I1203 11:14:04.988549 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a559e2e6-4fbc-4c51-99fb-5e971142d41e","Type":"ContainerDied","Data":"487f9b5d94f2404bbca651f0c2cec97ba35b47e2bfa264b0d4278db7e96f9e9a"} Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.001753 4646 scope.go:117] "RemoveContainer" containerID="d1dcde71ba498ce91f9efa7ca9dfefc43ef6daada1b6c97a966650feb4e99fd8" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.013597 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.020486 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.024065 4646 scope.go:117] "RemoveContainer" containerID="36bf63d44d74477babbf166dbefd3259be02a534d8b1c98399f5d279e99b6fcb" Dec 03 11:14:05 crc kubenswrapper[4646]: E1203 11:14:05.025069 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36bf63d44d74477babbf166dbefd3259be02a534d8b1c98399f5d279e99b6fcb\": container with ID starting with 36bf63d44d74477babbf166dbefd3259be02a534d8b1c98399f5d279e99b6fcb not found: ID does not exist" containerID="36bf63d44d74477babbf166dbefd3259be02a534d8b1c98399f5d279e99b6fcb" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.025123 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36bf63d44d74477babbf166dbefd3259be02a534d8b1c98399f5d279e99b6fcb"} err="failed to get container status \"36bf63d44d74477babbf166dbefd3259be02a534d8b1c98399f5d279e99b6fcb\": rpc error: code = NotFound desc = could not find container \"36bf63d44d74477babbf166dbefd3259be02a534d8b1c98399f5d279e99b6fcb\": container with ID starting with 36bf63d44d74477babbf166dbefd3259be02a534d8b1c98399f5d279e99b6fcb not found: ID does not exist" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.025150 4646 scope.go:117] "RemoveContainer" containerID="d1dcde71ba498ce91f9efa7ca9dfefc43ef6daada1b6c97a966650feb4e99fd8" Dec 03 11:14:05 crc kubenswrapper[4646]: E1203 11:14:05.025588 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1dcde71ba498ce91f9efa7ca9dfefc43ef6daada1b6c97a966650feb4e99fd8\": container with ID starting with d1dcde71ba498ce91f9efa7ca9dfefc43ef6daada1b6c97a966650feb4e99fd8 not found: ID does not exist" containerID="d1dcde71ba498ce91f9efa7ca9dfefc43ef6daada1b6c97a966650feb4e99fd8" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.025609 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1dcde71ba498ce91f9efa7ca9dfefc43ef6daada1b6c97a966650feb4e99fd8"} err="failed to get container status \"d1dcde71ba498ce91f9efa7ca9dfefc43ef6daada1b6c97a966650feb4e99fd8\": rpc error: code = NotFound desc = could not find container \"d1dcde71ba498ce91f9efa7ca9dfefc43ef6daada1b6c97a966650feb4e99fd8\": container with ID starting with d1dcde71ba498ce91f9efa7ca9dfefc43ef6daada1b6c97a966650feb4e99fd8 not found: ID does not exist" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.031735 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 11:14:05 crc kubenswrapper[4646]: E1203 11:14:05.032280 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a559e2e6-4fbc-4c51-99fb-5e971142d41e" containerName="cinder-api-log" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.032363 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="a559e2e6-4fbc-4c51-99fb-5e971142d41e" containerName="cinder-api-log" Dec 03 11:14:05 crc kubenswrapper[4646]: E1203 11:14:05.032492 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a559e2e6-4fbc-4c51-99fb-5e971142d41e" containerName="cinder-api" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.032573 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="a559e2e6-4fbc-4c51-99fb-5e971142d41e" containerName="cinder-api" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.032838 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="a559e2e6-4fbc-4c51-99fb-5e971142d41e" containerName="cinder-api" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.032933 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="a559e2e6-4fbc-4c51-99fb-5e971142d41e" containerName="cinder-api-log" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.033976 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.044674 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.044875 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.044894 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.050015 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.130417 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.130491 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-scripts\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.130576 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ttjh\" (UniqueName: \"kubernetes.io/projected/ad772e53-d750-4f6c-99be-ff551ebd0669-kube-api-access-9ttjh\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.130598 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.130638 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-config-data\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.130654 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ad772e53-d750-4f6c-99be-ff551ebd0669-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.130669 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-config-data-custom\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.130686 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.130703 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad772e53-d750-4f6c-99be-ff551ebd0669-logs\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.231500 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ttjh\" (UniqueName: \"kubernetes.io/projected/ad772e53-d750-4f6c-99be-ff551ebd0669-kube-api-access-9ttjh\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.231536 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.231572 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-config-data\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.231590 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ad772e53-d750-4f6c-99be-ff551ebd0669-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.231604 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-config-data-custom\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.231623 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.231639 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad772e53-d750-4f6c-99be-ff551ebd0669-logs\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.231655 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.231689 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-scripts\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.232849 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ad772e53-d750-4f6c-99be-ff551ebd0669-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.233149 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad772e53-d750-4f6c-99be-ff551ebd0669-logs\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.237954 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-scripts\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.238080 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.240948 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.243704 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.243825 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-config-data-custom\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.246229 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad772e53-d750-4f6c-99be-ff551ebd0669-config-data\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.254816 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ttjh\" (UniqueName: \"kubernetes.io/projected/ad772e53-d750-4f6c-99be-ff551ebd0669-kube-api-access-9ttjh\") pod \"cinder-api-0\" (UID: \"ad772e53-d750-4f6c-99be-ff551ebd0669\") " pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.364050 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.834210 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 11:14:05 crc kubenswrapper[4646]: W1203 11:14:05.845666 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad772e53_d750_4f6c_99be_ff551ebd0669.slice/crio-1dec62329b7af757c56848dac7ebe2744d71c5e912618fe67b8d99fb523418fc WatchSource:0}: Error finding container 1dec62329b7af757c56848dac7ebe2744d71c5e912618fe67b8d99fb523418fc: Status 404 returned error can't find the container with id 1dec62329b7af757c56848dac7ebe2744d71c5e912618fe67b8d99fb523418fc Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.869450 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a559e2e6-4fbc-4c51-99fb-5e971142d41e" path="/var/lib/kubelet/pods/a559e2e6-4fbc-4c51-99fb-5e971142d41e/volumes" Dec 03 11:14:05 crc kubenswrapper[4646]: I1203 11:14:05.995357 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"222c4fa4-6552-4021-ab4d-58f1a343a97d","Type":"ContainerStarted","Data":"bc87fdde229c087f2ef874ed2476c41c44e7bff5c06b36da90a4bb8b23a9ca0d"} Dec 03 11:14:06 crc kubenswrapper[4646]: I1203 11:14:06.002085 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ad772e53-d750-4f6c-99be-ff551ebd0669","Type":"ContainerStarted","Data":"1dec62329b7af757c56848dac7ebe2744d71c5e912618fe67b8d99fb523418fc"} Dec 03 11:14:07 crc kubenswrapper[4646]: I1203 11:14:07.015638 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"222c4fa4-6552-4021-ab4d-58f1a343a97d","Type":"ContainerStarted","Data":"264d31f64069361de409d56af8d37f8a67480a42d8e46d2b212d9161f0fa8c6a"} Dec 03 11:14:07 crc kubenswrapper[4646]: I1203 11:14:07.015799 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="ceilometer-central-agent" containerID="cri-o://92a29e17bde562ccf50740b4353f58418d356746e4b412f777f901691a62a737" gracePeriod=30 Dec 03 11:14:07 crc kubenswrapper[4646]: I1203 11:14:07.015838 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="sg-core" containerID="cri-o://bc87fdde229c087f2ef874ed2476c41c44e7bff5c06b36da90a4bb8b23a9ca0d" gracePeriod=30 Dec 03 11:14:07 crc kubenswrapper[4646]: I1203 11:14:07.015853 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="ceilometer-notification-agent" containerID="cri-o://e8f098272126786a2cdc6fa03c7b634611a0e63b9ae1517186f739a749575709" gracePeriod=30 Dec 03 11:14:07 crc kubenswrapper[4646]: I1203 11:14:07.015874 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="proxy-httpd" containerID="cri-o://264d31f64069361de409d56af8d37f8a67480a42d8e46d2b212d9161f0fa8c6a" gracePeriod=30 Dec 03 11:14:07 crc kubenswrapper[4646]: I1203 11:14:07.017633 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 11:14:07 crc kubenswrapper[4646]: I1203 11:14:07.037235 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.949062301 podStartE2EDuration="8.037220556s" podCreationTimestamp="2025-12-03 11:13:59 +0000 UTC" firstStartedPulling="2025-12-03 11:14:00.042741906 +0000 UTC m=+1216.505798041" lastFinishedPulling="2025-12-03 11:14:06.130900171 +0000 UTC m=+1222.593956296" observedRunningTime="2025-12-03 11:14:07.036683041 +0000 UTC m=+1223.499739186" watchObservedRunningTime="2025-12-03 11:14:07.037220556 +0000 UTC m=+1223.500276691" Dec 03 11:14:07 crc kubenswrapper[4646]: I1203 11:14:07.052072 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ad772e53-d750-4f6c-99be-ff551ebd0669","Type":"ContainerStarted","Data":"bf60fa39c5e378976fb276f42466f8df30c52c865f19ac06778e28614d74148b"} Dec 03 11:14:08 crc kubenswrapper[4646]: I1203 11:14:08.061260 4646 generic.go:334] "Generic (PLEG): container finished" podID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerID="264d31f64069361de409d56af8d37f8a67480a42d8e46d2b212d9161f0fa8c6a" exitCode=0 Dec 03 11:14:08 crc kubenswrapper[4646]: I1203 11:14:08.061630 4646 generic.go:334] "Generic (PLEG): container finished" podID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerID="bc87fdde229c087f2ef874ed2476c41c44e7bff5c06b36da90a4bb8b23a9ca0d" exitCode=2 Dec 03 11:14:08 crc kubenswrapper[4646]: I1203 11:14:08.061350 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"222c4fa4-6552-4021-ab4d-58f1a343a97d","Type":"ContainerDied","Data":"264d31f64069361de409d56af8d37f8a67480a42d8e46d2b212d9161f0fa8c6a"} Dec 03 11:14:08 crc kubenswrapper[4646]: I1203 11:14:08.061646 4646 generic.go:334] "Generic (PLEG): container finished" podID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerID="e8f098272126786a2cdc6fa03c7b634611a0e63b9ae1517186f739a749575709" exitCode=0 Dec 03 11:14:08 crc kubenswrapper[4646]: I1203 11:14:08.061674 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"222c4fa4-6552-4021-ab4d-58f1a343a97d","Type":"ContainerDied","Data":"bc87fdde229c087f2ef874ed2476c41c44e7bff5c06b36da90a4bb8b23a9ca0d"} Dec 03 11:14:08 crc kubenswrapper[4646]: I1203 11:14:08.061742 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"222c4fa4-6552-4021-ab4d-58f1a343a97d","Type":"ContainerDied","Data":"e8f098272126786a2cdc6fa03c7b634611a0e63b9ae1517186f739a749575709"} Dec 03 11:14:08 crc kubenswrapper[4646]: I1203 11:14:08.063285 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ad772e53-d750-4f6c-99be-ff551ebd0669","Type":"ContainerStarted","Data":"a6c2335ee034ec2774607180e407441a4a616796ebf2e2445eca8d614ffcd353"} Dec 03 11:14:08 crc kubenswrapper[4646]: I1203 11:14:08.063445 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 11:14:08 crc kubenswrapper[4646]: I1203 11:14:08.081448 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.081428206 podStartE2EDuration="3.081428206s" podCreationTimestamp="2025-12-03 11:14:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:14:08.080979724 +0000 UTC m=+1224.544035859" watchObservedRunningTime="2025-12-03 11:14:08.081428206 +0000 UTC m=+1224.544484341" Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.100702 4646 generic.go:334] "Generic (PLEG): container finished" podID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerID="92a29e17bde562ccf50740b4353f58418d356746e4b412f777f901691a62a737" exitCode=0 Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.100739 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"222c4fa4-6552-4021-ab4d-58f1a343a97d","Type":"ContainerDied","Data":"92a29e17bde562ccf50740b4353f58418d356746e4b412f777f901691a62a737"} Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.291626 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.327161 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-combined-ca-bundle\") pod \"222c4fa4-6552-4021-ab4d-58f1a343a97d\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.327220 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/222c4fa4-6552-4021-ab4d-58f1a343a97d-run-httpd\") pod \"222c4fa4-6552-4021-ab4d-58f1a343a97d\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.327370 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-scripts\") pod \"222c4fa4-6552-4021-ab4d-58f1a343a97d\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.327400 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-config-data\") pod \"222c4fa4-6552-4021-ab4d-58f1a343a97d\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.327457 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/222c4fa4-6552-4021-ab4d-58f1a343a97d-log-httpd\") pod \"222c4fa4-6552-4021-ab4d-58f1a343a97d\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.327487 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sztzw\" (UniqueName: \"kubernetes.io/projected/222c4fa4-6552-4021-ab4d-58f1a343a97d-kube-api-access-sztzw\") pod \"222c4fa4-6552-4021-ab4d-58f1a343a97d\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.327536 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-sg-core-conf-yaml\") pod \"222c4fa4-6552-4021-ab4d-58f1a343a97d\" (UID: \"222c4fa4-6552-4021-ab4d-58f1a343a97d\") " Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.338624 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/222c4fa4-6552-4021-ab4d-58f1a343a97d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "222c4fa4-6552-4021-ab4d-58f1a343a97d" (UID: "222c4fa4-6552-4021-ab4d-58f1a343a97d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.339000 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/222c4fa4-6552-4021-ab4d-58f1a343a97d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "222c4fa4-6552-4021-ab4d-58f1a343a97d" (UID: "222c4fa4-6552-4021-ab4d-58f1a343a97d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.342723 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-scripts" (OuterVolumeSpecName: "scripts") pod "222c4fa4-6552-4021-ab4d-58f1a343a97d" (UID: "222c4fa4-6552-4021-ab4d-58f1a343a97d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.365543 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/222c4fa4-6552-4021-ab4d-58f1a343a97d-kube-api-access-sztzw" (OuterVolumeSpecName: "kube-api-access-sztzw") pod "222c4fa4-6552-4021-ab4d-58f1a343a97d" (UID: "222c4fa4-6552-4021-ab4d-58f1a343a97d"). InnerVolumeSpecName "kube-api-access-sztzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.377592 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "222c4fa4-6552-4021-ab4d-58f1a343a97d" (UID: "222c4fa4-6552-4021-ab4d-58f1a343a97d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.431542 4646 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.431582 4646 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/222c4fa4-6552-4021-ab4d-58f1a343a97d-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.431593 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.431604 4646 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/222c4fa4-6552-4021-ab4d-58f1a343a97d-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.431617 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sztzw\" (UniqueName: \"kubernetes.io/projected/222c4fa4-6552-4021-ab4d-58f1a343a97d-kube-api-access-sztzw\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.448152 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "222c4fa4-6552-4021-ab4d-58f1a343a97d" (UID: "222c4fa4-6552-4021-ab4d-58f1a343a97d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.492094 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-config-data" (OuterVolumeSpecName: "config-data") pod "222c4fa4-6552-4021-ab4d-58f1a343a97d" (UID: "222c4fa4-6552-4021-ab4d-58f1a343a97d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.532771 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:10 crc kubenswrapper[4646]: I1203 11:14:10.532808 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/222c4fa4-6552-4021-ab4d-58f1a343a97d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.112406 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"222c4fa4-6552-4021-ab4d-58f1a343a97d","Type":"ContainerDied","Data":"dab4f562a5dae6a79960c8fef75aad4b44c8af54f272ea14878b5617db65164a"} Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.112465 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.112800 4646 scope.go:117] "RemoveContainer" containerID="264d31f64069361de409d56af8d37f8a67480a42d8e46d2b212d9161f0fa8c6a" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.133145 4646 scope.go:117] "RemoveContainer" containerID="bc87fdde229c087f2ef874ed2476c41c44e7bff5c06b36da90a4bb8b23a9ca0d" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.150843 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.158314 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.167522 4646 scope.go:117] "RemoveContainer" containerID="e8f098272126786a2cdc6fa03c7b634611a0e63b9ae1517186f739a749575709" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.184024 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:14:11 crc kubenswrapper[4646]: E1203 11:14:11.184440 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="ceilometer-notification-agent" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.184456 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="ceilometer-notification-agent" Dec 03 11:14:11 crc kubenswrapper[4646]: E1203 11:14:11.184480 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="ceilometer-central-agent" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.184487 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="ceilometer-central-agent" Dec 03 11:14:11 crc kubenswrapper[4646]: E1203 11:14:11.184500 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="proxy-httpd" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.184507 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="proxy-httpd" Dec 03 11:14:11 crc kubenswrapper[4646]: E1203 11:14:11.184524 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="sg-core" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.184530 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="sg-core" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.185468 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="ceilometer-central-agent" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.185490 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="proxy-httpd" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.185503 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="ceilometer-notification-agent" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.185516 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" containerName="sg-core" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.187491 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.192166 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.192434 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.193611 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.210393 4646 scope.go:117] "RemoveContainer" containerID="92a29e17bde562ccf50740b4353f58418d356746e4b412f777f901691a62a737" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.247837 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpcxs\" (UniqueName: \"kubernetes.io/projected/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-kube-api-access-vpcxs\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.247900 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-log-httpd\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.247920 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-run-httpd\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.247964 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-scripts\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.247984 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.248011 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.248037 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-config-data\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.350002 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-log-httpd\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.350243 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-run-httpd\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.350406 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-scripts\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.350542 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.350670 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.350777 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-config-data\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.350800 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-log-httpd\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.350810 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-run-httpd\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.350959 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpcxs\" (UniqueName: \"kubernetes.io/projected/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-kube-api-access-vpcxs\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.354669 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.354838 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-config-data\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.360740 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.363852 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-scripts\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.368151 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpcxs\" (UniqueName: \"kubernetes.io/projected/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-kube-api-access-vpcxs\") pod \"ceilometer-0\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.512292 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.861744 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="222c4fa4-6552-4021-ab4d-58f1a343a97d" path="/var/lib/kubelet/pods/222c4fa4-6552-4021-ab4d-58f1a343a97d/volumes" Dec 03 11:14:11 crc kubenswrapper[4646]: I1203 11:14:11.985752 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:14:12 crc kubenswrapper[4646]: I1203 11:14:12.121108 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16532a7f-0c9f-4a66-83b0-a3f1d731ac94","Type":"ContainerStarted","Data":"28cc20390276c58b8adb94f7aad3a4f4d611f0bac6d444868ced3fe707401270"} Dec 03 11:14:14 crc kubenswrapper[4646]: I1203 11:14:14.140990 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16532a7f-0c9f-4a66-83b0-a3f1d731ac94","Type":"ContainerStarted","Data":"9d3f3008df0625fdf188614b4f01c299dab949acd190092206d2244869a11277"} Dec 03 11:14:15 crc kubenswrapper[4646]: I1203 11:14:15.173082 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16532a7f-0c9f-4a66-83b0-a3f1d731ac94","Type":"ContainerStarted","Data":"0c97fe281a42412bdce76422505dcb9eba9427ee9e31d63d8f44a2c93fbfce1c"} Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.183009 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16532a7f-0c9f-4a66-83b0-a3f1d731ac94","Type":"ContainerStarted","Data":"123f7f2d16586d8ef519036a810613223a2a93969fcb5f188529c6296ddd255a"} Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.602913 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-4hgz5"] Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.604422 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4hgz5" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.624997 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4hgz5"] Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.658213 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l62tp\" (UniqueName: \"kubernetes.io/projected/21f00b80-abd7-49b1-8323-1f6d72a02455-kube-api-access-l62tp\") pod \"nova-api-db-create-4hgz5\" (UID: \"21f00b80-abd7-49b1-8323-1f6d72a02455\") " pod="openstack/nova-api-db-create-4hgz5" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.658271 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21f00b80-abd7-49b1-8323-1f6d72a02455-operator-scripts\") pod \"nova-api-db-create-4hgz5\" (UID: \"21f00b80-abd7-49b1-8323-1f6d72a02455\") " pod="openstack/nova-api-db-create-4hgz5" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.759311 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l62tp\" (UniqueName: \"kubernetes.io/projected/21f00b80-abd7-49b1-8323-1f6d72a02455-kube-api-access-l62tp\") pod \"nova-api-db-create-4hgz5\" (UID: \"21f00b80-abd7-49b1-8323-1f6d72a02455\") " pod="openstack/nova-api-db-create-4hgz5" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.759631 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21f00b80-abd7-49b1-8323-1f6d72a02455-operator-scripts\") pod \"nova-api-db-create-4hgz5\" (UID: \"21f00b80-abd7-49b1-8323-1f6d72a02455\") " pod="openstack/nova-api-db-create-4hgz5" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.760470 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21f00b80-abd7-49b1-8323-1f6d72a02455-operator-scripts\") pod \"nova-api-db-create-4hgz5\" (UID: \"21f00b80-abd7-49b1-8323-1f6d72a02455\") " pod="openstack/nova-api-db-create-4hgz5" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.771858 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-563f-account-create-update-kkpjv"] Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.773099 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-563f-account-create-update-kkpjv" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.785143 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-p45hd"] Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.786440 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p45hd" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.792763 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.798142 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-563f-account-create-update-kkpjv"] Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.808399 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-p45hd"] Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.822826 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l62tp\" (UniqueName: \"kubernetes.io/projected/21f00b80-abd7-49b1-8323-1f6d72a02455-kube-api-access-l62tp\") pod \"nova-api-db-create-4hgz5\" (UID: \"21f00b80-abd7-49b1-8323-1f6d72a02455\") " pod="openstack/nova-api-db-create-4hgz5" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.863528 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9qdd\" (UniqueName: \"kubernetes.io/projected/2fda3961-30ca-44c7-94f0-591558331b7d-kube-api-access-t9qdd\") pod \"nova-cell0-db-create-p45hd\" (UID: \"2fda3961-30ca-44c7-94f0-591558331b7d\") " pod="openstack/nova-cell0-db-create-p45hd" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.863619 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e229dd0-7f9f-4705-804d-584cbb33869c-operator-scripts\") pod \"nova-api-563f-account-create-update-kkpjv\" (UID: \"3e229dd0-7f9f-4705-804d-584cbb33869c\") " pod="openstack/nova-api-563f-account-create-update-kkpjv" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.863675 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd4g7\" (UniqueName: \"kubernetes.io/projected/3e229dd0-7f9f-4705-804d-584cbb33869c-kube-api-access-cd4g7\") pod \"nova-api-563f-account-create-update-kkpjv\" (UID: \"3e229dd0-7f9f-4705-804d-584cbb33869c\") " pod="openstack/nova-api-563f-account-create-update-kkpjv" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.863711 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fda3961-30ca-44c7-94f0-591558331b7d-operator-scripts\") pod \"nova-cell0-db-create-p45hd\" (UID: \"2fda3961-30ca-44c7-94f0-591558331b7d\") " pod="openstack/nova-cell0-db-create-p45hd" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.915390 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-jrbcg"] Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.916793 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jrbcg" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.926573 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jrbcg"] Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.965685 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8adb3112-e79b-40a3-a906-65bd37489c7a-operator-scripts\") pod \"nova-cell1-db-create-jrbcg\" (UID: \"8adb3112-e79b-40a3-a906-65bd37489c7a\") " pod="openstack/nova-cell1-db-create-jrbcg" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.966036 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x476n\" (UniqueName: \"kubernetes.io/projected/8adb3112-e79b-40a3-a906-65bd37489c7a-kube-api-access-x476n\") pod \"nova-cell1-db-create-jrbcg\" (UID: \"8adb3112-e79b-40a3-a906-65bd37489c7a\") " pod="openstack/nova-cell1-db-create-jrbcg" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.966192 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9qdd\" (UniqueName: \"kubernetes.io/projected/2fda3961-30ca-44c7-94f0-591558331b7d-kube-api-access-t9qdd\") pod \"nova-cell0-db-create-p45hd\" (UID: \"2fda3961-30ca-44c7-94f0-591558331b7d\") " pod="openstack/nova-cell0-db-create-p45hd" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.966368 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e229dd0-7f9f-4705-804d-584cbb33869c-operator-scripts\") pod \"nova-api-563f-account-create-update-kkpjv\" (UID: \"3e229dd0-7f9f-4705-804d-584cbb33869c\") " pod="openstack/nova-api-563f-account-create-update-kkpjv" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.966542 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd4g7\" (UniqueName: \"kubernetes.io/projected/3e229dd0-7f9f-4705-804d-584cbb33869c-kube-api-access-cd4g7\") pod \"nova-api-563f-account-create-update-kkpjv\" (UID: \"3e229dd0-7f9f-4705-804d-584cbb33869c\") " pod="openstack/nova-api-563f-account-create-update-kkpjv" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.966671 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fda3961-30ca-44c7-94f0-591558331b7d-operator-scripts\") pod \"nova-cell0-db-create-p45hd\" (UID: \"2fda3961-30ca-44c7-94f0-591558331b7d\") " pod="openstack/nova-cell0-db-create-p45hd" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.971517 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e229dd0-7f9f-4705-804d-584cbb33869c-operator-scripts\") pod \"nova-api-563f-account-create-update-kkpjv\" (UID: \"3e229dd0-7f9f-4705-804d-584cbb33869c\") " pod="openstack/nova-api-563f-account-create-update-kkpjv" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.977259 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fda3961-30ca-44c7-94f0-591558331b7d-operator-scripts\") pod \"nova-cell0-db-create-p45hd\" (UID: \"2fda3961-30ca-44c7-94f0-591558331b7d\") " pod="openstack/nova-cell0-db-create-p45hd" Dec 03 11:14:16 crc kubenswrapper[4646]: I1203 11:14:16.992850 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9qdd\" (UniqueName: \"kubernetes.io/projected/2fda3961-30ca-44c7-94f0-591558331b7d-kube-api-access-t9qdd\") pod \"nova-cell0-db-create-p45hd\" (UID: \"2fda3961-30ca-44c7-94f0-591558331b7d\") " pod="openstack/nova-cell0-db-create-p45hd" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.003366 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd4g7\" (UniqueName: \"kubernetes.io/projected/3e229dd0-7f9f-4705-804d-584cbb33869c-kube-api-access-cd4g7\") pod \"nova-api-563f-account-create-update-kkpjv\" (UID: \"3e229dd0-7f9f-4705-804d-584cbb33869c\") " pod="openstack/nova-api-563f-account-create-update-kkpjv" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.013143 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4hgz5" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.061398 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-7483-account-create-update-78pjg"] Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.062614 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7483-account-create-update-78pjg" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.066786 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.068252 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x476n\" (UniqueName: \"kubernetes.io/projected/8adb3112-e79b-40a3-a906-65bd37489c7a-kube-api-access-x476n\") pod \"nova-cell1-db-create-jrbcg\" (UID: \"8adb3112-e79b-40a3-a906-65bd37489c7a\") " pod="openstack/nova-cell1-db-create-jrbcg" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.068566 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8adb3112-e79b-40a3-a906-65bd37489c7a-operator-scripts\") pod \"nova-cell1-db-create-jrbcg\" (UID: \"8adb3112-e79b-40a3-a906-65bd37489c7a\") " pod="openstack/nova-cell1-db-create-jrbcg" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.069423 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8adb3112-e79b-40a3-a906-65bd37489c7a-operator-scripts\") pod \"nova-cell1-db-create-jrbcg\" (UID: \"8adb3112-e79b-40a3-a906-65bd37489c7a\") " pod="openstack/nova-cell1-db-create-jrbcg" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.090643 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x476n\" (UniqueName: \"kubernetes.io/projected/8adb3112-e79b-40a3-a906-65bd37489c7a-kube-api-access-x476n\") pod \"nova-cell1-db-create-jrbcg\" (UID: \"8adb3112-e79b-40a3-a906-65bd37489c7a\") " pod="openstack/nova-cell1-db-create-jrbcg" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.128042 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7483-account-create-update-78pjg"] Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.170189 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p45hd" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.170576 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-563f-account-create-update-kkpjv" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.172450 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/406832ee-9a27-4930-a47e-0fc8cf4b3a86-operator-scripts\") pod \"nova-cell0-7483-account-create-update-78pjg\" (UID: \"406832ee-9a27-4930-a47e-0fc8cf4b3a86\") " pod="openstack/nova-cell0-7483-account-create-update-78pjg" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.172588 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctpnw\" (UniqueName: \"kubernetes.io/projected/406832ee-9a27-4930-a47e-0fc8cf4b3a86-kube-api-access-ctpnw\") pod \"nova-cell0-7483-account-create-update-78pjg\" (UID: \"406832ee-9a27-4930-a47e-0fc8cf4b3a86\") " pod="openstack/nova-cell0-7483-account-create-update-78pjg" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.213087 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-1990-account-create-update-tc486"] Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.216258 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1990-account-create-update-tc486" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.218680 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16532a7f-0c9f-4a66-83b0-a3f1d731ac94","Type":"ContainerStarted","Data":"4192479ffb79202d352ccc6c1693bdb5c0524f53d626432908e0f22fadc730f3"} Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.219048 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.219504 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.249105 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jrbcg" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.249262 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1990-account-create-update-tc486"] Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.269130 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.764806309 podStartE2EDuration="6.269112244s" podCreationTimestamp="2025-12-03 11:14:11 +0000 UTC" firstStartedPulling="2025-12-03 11:14:11.99399766 +0000 UTC m=+1228.457053795" lastFinishedPulling="2025-12-03 11:14:16.498303595 +0000 UTC m=+1232.961359730" observedRunningTime="2025-12-03 11:14:17.257479876 +0000 UTC m=+1233.720536011" watchObservedRunningTime="2025-12-03 11:14:17.269112244 +0000 UTC m=+1233.732168379" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.275668 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctpnw\" (UniqueName: \"kubernetes.io/projected/406832ee-9a27-4930-a47e-0fc8cf4b3a86-kube-api-access-ctpnw\") pod \"nova-cell0-7483-account-create-update-78pjg\" (UID: \"406832ee-9a27-4930-a47e-0fc8cf4b3a86\") " pod="openstack/nova-cell0-7483-account-create-update-78pjg" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.275748 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/406832ee-9a27-4930-a47e-0fc8cf4b3a86-operator-scripts\") pod \"nova-cell0-7483-account-create-update-78pjg\" (UID: \"406832ee-9a27-4930-a47e-0fc8cf4b3a86\") " pod="openstack/nova-cell0-7483-account-create-update-78pjg" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.276692 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/406832ee-9a27-4930-a47e-0fc8cf4b3a86-operator-scripts\") pod \"nova-cell0-7483-account-create-update-78pjg\" (UID: \"406832ee-9a27-4930-a47e-0fc8cf4b3a86\") " pod="openstack/nova-cell0-7483-account-create-update-78pjg" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.293907 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctpnw\" (UniqueName: \"kubernetes.io/projected/406832ee-9a27-4930-a47e-0fc8cf4b3a86-kube-api-access-ctpnw\") pod \"nova-cell0-7483-account-create-update-78pjg\" (UID: \"406832ee-9a27-4930-a47e-0fc8cf4b3a86\") " pod="openstack/nova-cell0-7483-account-create-update-78pjg" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.383327 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc2pt\" (UniqueName: \"kubernetes.io/projected/cfd139ae-cd72-4d79-9fd2-ad5f199a33be-kube-api-access-dc2pt\") pod \"nova-cell1-1990-account-create-update-tc486\" (UID: \"cfd139ae-cd72-4d79-9fd2-ad5f199a33be\") " pod="openstack/nova-cell1-1990-account-create-update-tc486" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.383755 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cfd139ae-cd72-4d79-9fd2-ad5f199a33be-operator-scripts\") pod \"nova-cell1-1990-account-create-update-tc486\" (UID: \"cfd139ae-cd72-4d79-9fd2-ad5f199a33be\") " pod="openstack/nova-cell1-1990-account-create-update-tc486" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.384980 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7483-account-create-update-78pjg" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.485619 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc2pt\" (UniqueName: \"kubernetes.io/projected/cfd139ae-cd72-4d79-9fd2-ad5f199a33be-kube-api-access-dc2pt\") pod \"nova-cell1-1990-account-create-update-tc486\" (UID: \"cfd139ae-cd72-4d79-9fd2-ad5f199a33be\") " pod="openstack/nova-cell1-1990-account-create-update-tc486" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.485684 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cfd139ae-cd72-4d79-9fd2-ad5f199a33be-operator-scripts\") pod \"nova-cell1-1990-account-create-update-tc486\" (UID: \"cfd139ae-cd72-4d79-9fd2-ad5f199a33be\") " pod="openstack/nova-cell1-1990-account-create-update-tc486" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.486418 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cfd139ae-cd72-4d79-9fd2-ad5f199a33be-operator-scripts\") pod \"nova-cell1-1990-account-create-update-tc486\" (UID: \"cfd139ae-cd72-4d79-9fd2-ad5f199a33be\") " pod="openstack/nova-cell1-1990-account-create-update-tc486" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.536927 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc2pt\" (UniqueName: \"kubernetes.io/projected/cfd139ae-cd72-4d79-9fd2-ad5f199a33be-kube-api-access-dc2pt\") pod \"nova-cell1-1990-account-create-update-tc486\" (UID: \"cfd139ae-cd72-4d79-9fd2-ad5f199a33be\") " pod="openstack/nova-cell1-1990-account-create-update-tc486" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.548962 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1990-account-create-update-tc486" Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.715185 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4hgz5"] Dec 03 11:14:17 crc kubenswrapper[4646]: I1203 11:14:17.920491 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-p45hd"] Dec 03 11:14:18 crc kubenswrapper[4646]: I1203 11:14:18.005706 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jrbcg"] Dec 03 11:14:18 crc kubenswrapper[4646]: I1203 11:14:18.033090 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-563f-account-create-update-kkpjv"] Dec 03 11:14:18 crc kubenswrapper[4646]: I1203 11:14:18.226966 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7483-account-create-update-78pjg"] Dec 03 11:14:18 crc kubenswrapper[4646]: I1203 11:14:18.247174 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-563f-account-create-update-kkpjv" event={"ID":"3e229dd0-7f9f-4705-804d-584cbb33869c","Type":"ContainerStarted","Data":"9f9dbe9326847da09585dc4f15fdf3a4f85fb5b8072fdb01c641c71fb53b3379"} Dec 03 11:14:18 crc kubenswrapper[4646]: I1203 11:14:18.250090 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4hgz5" event={"ID":"21f00b80-abd7-49b1-8323-1f6d72a02455","Type":"ContainerStarted","Data":"e8f98fa462b3c9ef944c54836bfe32794de06a55d798726ecafd6123ec2ebfd5"} Dec 03 11:14:18 crc kubenswrapper[4646]: I1203 11:14:18.250136 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4hgz5" event={"ID":"21f00b80-abd7-49b1-8323-1f6d72a02455","Type":"ContainerStarted","Data":"1751a2427db5c5b4c2663bef83a9c98db3fae54b7a43b8d9b5bf35d5b716071c"} Dec 03 11:14:18 crc kubenswrapper[4646]: I1203 11:14:18.261535 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-p45hd" event={"ID":"2fda3961-30ca-44c7-94f0-591558331b7d","Type":"ContainerStarted","Data":"eb202eb66867be50b3daeb058d3496ac55c25d0ce20a3c3ee90edd628c599838"} Dec 03 11:14:18 crc kubenswrapper[4646]: I1203 11:14:18.261593 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-p45hd" event={"ID":"2fda3961-30ca-44c7-94f0-591558331b7d","Type":"ContainerStarted","Data":"98cd512b0a6a35d62115409a5084a89b5db12765fa2f00af40159c1cdbd5f535"} Dec 03 11:14:18 crc kubenswrapper[4646]: I1203 11:14:18.272247 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-4hgz5" podStartSLOduration=2.272226449 podStartE2EDuration="2.272226449s" podCreationTimestamp="2025-12-03 11:14:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:14:18.264276595 +0000 UTC m=+1234.727332730" watchObservedRunningTime="2025-12-03 11:14:18.272226449 +0000 UTC m=+1234.735282584" Dec 03 11:14:18 crc kubenswrapper[4646]: I1203 11:14:18.277820 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jrbcg" event={"ID":"8adb3112-e79b-40a3-a906-65bd37489c7a","Type":"ContainerStarted","Data":"87bb9e1a0fb97ff7fe08bc298a5f769dcad51ae3fb40c620339fe2aec11e0903"} Dec 03 11:14:18 crc kubenswrapper[4646]: I1203 11:14:18.286070 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-p45hd" podStartSLOduration=2.286051499 podStartE2EDuration="2.286051499s" podCreationTimestamp="2025-12-03 11:14:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:14:18.278608939 +0000 UTC m=+1234.741665074" watchObservedRunningTime="2025-12-03 11:14:18.286051499 +0000 UTC m=+1234.749107634" Dec 03 11:14:18 crc kubenswrapper[4646]: W1203 11:14:18.337858 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfd139ae_cd72_4d79_9fd2_ad5f199a33be.slice/crio-dc147a4ee5b613f594cb39d010bf24b6f80c8d1760ab19ab09f37f5ed66ac9e3 WatchSource:0}: Error finding container dc147a4ee5b613f594cb39d010bf24b6f80c8d1760ab19ab09f37f5ed66ac9e3: Status 404 returned error can't find the container with id dc147a4ee5b613f594cb39d010bf24b6f80c8d1760ab19ab09f37f5ed66ac9e3 Dec 03 11:14:18 crc kubenswrapper[4646]: I1203 11:14:18.339288 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1990-account-create-update-tc486"] Dec 03 11:14:18 crc kubenswrapper[4646]: I1203 11:14:18.808031 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 11:14:19 crc kubenswrapper[4646]: I1203 11:14:19.287510 4646 generic.go:334] "Generic (PLEG): container finished" podID="2fda3961-30ca-44c7-94f0-591558331b7d" containerID="eb202eb66867be50b3daeb058d3496ac55c25d0ce20a3c3ee90edd628c599838" exitCode=0 Dec 03 11:14:19 crc kubenswrapper[4646]: I1203 11:14:19.287588 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-p45hd" event={"ID":"2fda3961-30ca-44c7-94f0-591558331b7d","Type":"ContainerDied","Data":"eb202eb66867be50b3daeb058d3496ac55c25d0ce20a3c3ee90edd628c599838"} Dec 03 11:14:19 crc kubenswrapper[4646]: I1203 11:14:19.290348 4646 generic.go:334] "Generic (PLEG): container finished" podID="cfd139ae-cd72-4d79-9fd2-ad5f199a33be" containerID="c8e2e34e23acc940dc29e98773a2b9bdd9fd6838c95a147efd2e916d112542eb" exitCode=0 Dec 03 11:14:19 crc kubenswrapper[4646]: I1203 11:14:19.290738 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1990-account-create-update-tc486" event={"ID":"cfd139ae-cd72-4d79-9fd2-ad5f199a33be","Type":"ContainerDied","Data":"c8e2e34e23acc940dc29e98773a2b9bdd9fd6838c95a147efd2e916d112542eb"} Dec 03 11:14:19 crc kubenswrapper[4646]: I1203 11:14:19.290800 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1990-account-create-update-tc486" event={"ID":"cfd139ae-cd72-4d79-9fd2-ad5f199a33be","Type":"ContainerStarted","Data":"dc147a4ee5b613f594cb39d010bf24b6f80c8d1760ab19ab09f37f5ed66ac9e3"} Dec 03 11:14:19 crc kubenswrapper[4646]: I1203 11:14:19.291863 4646 generic.go:334] "Generic (PLEG): container finished" podID="8adb3112-e79b-40a3-a906-65bd37489c7a" containerID="977beaf3e97f2c5ba1aae4b5f12a9422009c9ce4eda726e04fc4d06dbe3a191b" exitCode=0 Dec 03 11:14:19 crc kubenswrapper[4646]: I1203 11:14:19.291922 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jrbcg" event={"ID":"8adb3112-e79b-40a3-a906-65bd37489c7a","Type":"ContainerDied","Data":"977beaf3e97f2c5ba1aae4b5f12a9422009c9ce4eda726e04fc4d06dbe3a191b"} Dec 03 11:14:19 crc kubenswrapper[4646]: I1203 11:14:19.293120 4646 generic.go:334] "Generic (PLEG): container finished" podID="406832ee-9a27-4930-a47e-0fc8cf4b3a86" containerID="c1173e9b04403cf852617ac87961202b9faed3962ab3ffe2e24b34617fd35f02" exitCode=0 Dec 03 11:14:19 crc kubenswrapper[4646]: I1203 11:14:19.293173 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7483-account-create-update-78pjg" event={"ID":"406832ee-9a27-4930-a47e-0fc8cf4b3a86","Type":"ContainerDied","Data":"c1173e9b04403cf852617ac87961202b9faed3962ab3ffe2e24b34617fd35f02"} Dec 03 11:14:19 crc kubenswrapper[4646]: I1203 11:14:19.293192 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7483-account-create-update-78pjg" event={"ID":"406832ee-9a27-4930-a47e-0fc8cf4b3a86","Type":"ContainerStarted","Data":"f2c0593a15943a174eade46635abb8ae668588b9a066571c6c564670ee142f2d"} Dec 03 11:14:19 crc kubenswrapper[4646]: I1203 11:14:19.294810 4646 generic.go:334] "Generic (PLEG): container finished" podID="3e229dd0-7f9f-4705-804d-584cbb33869c" containerID="e54fe6028631354001b72bf4442de6745834ddc33e3e554c3265045db9824897" exitCode=0 Dec 03 11:14:19 crc kubenswrapper[4646]: I1203 11:14:19.294872 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-563f-account-create-update-kkpjv" event={"ID":"3e229dd0-7f9f-4705-804d-584cbb33869c","Type":"ContainerDied","Data":"e54fe6028631354001b72bf4442de6745834ddc33e3e554c3265045db9824897"} Dec 03 11:14:19 crc kubenswrapper[4646]: I1203 11:14:19.296298 4646 generic.go:334] "Generic (PLEG): container finished" podID="21f00b80-abd7-49b1-8323-1f6d72a02455" containerID="e8f98fa462b3c9ef944c54836bfe32794de06a55d798726ecafd6123ec2ebfd5" exitCode=0 Dec 03 11:14:19 crc kubenswrapper[4646]: I1203 11:14:19.296365 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4hgz5" event={"ID":"21f00b80-abd7-49b1-8323-1f6d72a02455","Type":"ContainerDied","Data":"e8f98fa462b3c9ef944c54836bfe32794de06a55d798726ecafd6123ec2ebfd5"} Dec 03 11:14:20 crc kubenswrapper[4646]: I1203 11:14:20.933676 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jrbcg" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.025798 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x476n\" (UniqueName: \"kubernetes.io/projected/8adb3112-e79b-40a3-a906-65bd37489c7a-kube-api-access-x476n\") pod \"8adb3112-e79b-40a3-a906-65bd37489c7a\" (UID: \"8adb3112-e79b-40a3-a906-65bd37489c7a\") " Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.026080 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8adb3112-e79b-40a3-a906-65bd37489c7a-operator-scripts\") pod \"8adb3112-e79b-40a3-a906-65bd37489c7a\" (UID: \"8adb3112-e79b-40a3-a906-65bd37489c7a\") " Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.027027 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8adb3112-e79b-40a3-a906-65bd37489c7a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8adb3112-e79b-40a3-a906-65bd37489c7a" (UID: "8adb3112-e79b-40a3-a906-65bd37489c7a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.038523 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8adb3112-e79b-40a3-a906-65bd37489c7a-kube-api-access-x476n" (OuterVolumeSpecName: "kube-api-access-x476n") pod "8adb3112-e79b-40a3-a906-65bd37489c7a" (UID: "8adb3112-e79b-40a3-a906-65bd37489c7a"). InnerVolumeSpecName "kube-api-access-x476n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.127281 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8adb3112-e79b-40a3-a906-65bd37489c7a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.127322 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x476n\" (UniqueName: \"kubernetes.io/projected/8adb3112-e79b-40a3-a906-65bd37489c7a-kube-api-access-x476n\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.160184 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4hgz5" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.171426 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1990-account-create-update-tc486" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.185327 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-563f-account-create-update-kkpjv" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.197805 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7483-account-create-update-78pjg" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.205620 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p45hd" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.314090 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1990-account-create-update-tc486" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.314763 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1990-account-create-update-tc486" event={"ID":"cfd139ae-cd72-4d79-9fd2-ad5f199a33be","Type":"ContainerDied","Data":"dc147a4ee5b613f594cb39d010bf24b6f80c8d1760ab19ab09f37f5ed66ac9e3"} Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.314872 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc147a4ee5b613f594cb39d010bf24b6f80c8d1760ab19ab09f37f5ed66ac9e3" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.316446 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jrbcg" event={"ID":"8adb3112-e79b-40a3-a906-65bd37489c7a","Type":"ContainerDied","Data":"87bb9e1a0fb97ff7fe08bc298a5f769dcad51ae3fb40c620339fe2aec11e0903"} Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.316487 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87bb9e1a0fb97ff7fe08bc298a5f769dcad51ae3fb40c620339fe2aec11e0903" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.316422 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jrbcg" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.318082 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7483-account-create-update-78pjg" event={"ID":"406832ee-9a27-4930-a47e-0fc8cf4b3a86","Type":"ContainerDied","Data":"f2c0593a15943a174eade46635abb8ae668588b9a066571c6c564670ee142f2d"} Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.318196 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2c0593a15943a174eade46635abb8ae668588b9a066571c6c564670ee142f2d" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.318120 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7483-account-create-update-78pjg" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.319346 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-563f-account-create-update-kkpjv" event={"ID":"3e229dd0-7f9f-4705-804d-584cbb33869c","Type":"ContainerDied","Data":"9f9dbe9326847da09585dc4f15fdf3a4f85fb5b8072fdb01c641c71fb53b3379"} Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.319381 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f9dbe9326847da09585dc4f15fdf3a4f85fb5b8072fdb01c641c71fb53b3379" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.319440 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-563f-account-create-update-kkpjv" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.321171 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4hgz5" event={"ID":"21f00b80-abd7-49b1-8323-1f6d72a02455","Type":"ContainerDied","Data":"1751a2427db5c5b4c2663bef83a9c98db3fae54b7a43b8d9b5bf35d5b716071c"} Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.321259 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1751a2427db5c5b4c2663bef83a9c98db3fae54b7a43b8d9b5bf35d5b716071c" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.321186 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4hgz5" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.322919 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p45hd" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.322940 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-p45hd" event={"ID":"2fda3961-30ca-44c7-94f0-591558331b7d","Type":"ContainerDied","Data":"98cd512b0a6a35d62115409a5084a89b5db12765fa2f00af40159c1cdbd5f535"} Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.323084 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98cd512b0a6a35d62115409a5084a89b5db12765fa2f00af40159c1cdbd5f535" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.344281 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9qdd\" (UniqueName: \"kubernetes.io/projected/2fda3961-30ca-44c7-94f0-591558331b7d-kube-api-access-t9qdd\") pod \"2fda3961-30ca-44c7-94f0-591558331b7d\" (UID: \"2fda3961-30ca-44c7-94f0-591558331b7d\") " Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.344466 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fda3961-30ca-44c7-94f0-591558331b7d-operator-scripts\") pod \"2fda3961-30ca-44c7-94f0-591558331b7d\" (UID: \"2fda3961-30ca-44c7-94f0-591558331b7d\") " Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.344637 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cfd139ae-cd72-4d79-9fd2-ad5f199a33be-operator-scripts\") pod \"cfd139ae-cd72-4d79-9fd2-ad5f199a33be\" (UID: \"cfd139ae-cd72-4d79-9fd2-ad5f199a33be\") " Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.344750 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/406832ee-9a27-4930-a47e-0fc8cf4b3a86-operator-scripts\") pod \"406832ee-9a27-4930-a47e-0fc8cf4b3a86\" (UID: \"406832ee-9a27-4930-a47e-0fc8cf4b3a86\") " Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.344846 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l62tp\" (UniqueName: \"kubernetes.io/projected/21f00b80-abd7-49b1-8323-1f6d72a02455-kube-api-access-l62tp\") pod \"21f00b80-abd7-49b1-8323-1f6d72a02455\" (UID: \"21f00b80-abd7-49b1-8323-1f6d72a02455\") " Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.344988 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e229dd0-7f9f-4705-804d-584cbb33869c-operator-scripts\") pod \"3e229dd0-7f9f-4705-804d-584cbb33869c\" (UID: \"3e229dd0-7f9f-4705-804d-584cbb33869c\") " Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.345134 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21f00b80-abd7-49b1-8323-1f6d72a02455-operator-scripts\") pod \"21f00b80-abd7-49b1-8323-1f6d72a02455\" (UID: \"21f00b80-abd7-49b1-8323-1f6d72a02455\") " Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.345222 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd4g7\" (UniqueName: \"kubernetes.io/projected/3e229dd0-7f9f-4705-804d-584cbb33869c-kube-api-access-cd4g7\") pod \"3e229dd0-7f9f-4705-804d-584cbb33869c\" (UID: \"3e229dd0-7f9f-4705-804d-584cbb33869c\") " Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.345286 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctpnw\" (UniqueName: \"kubernetes.io/projected/406832ee-9a27-4930-a47e-0fc8cf4b3a86-kube-api-access-ctpnw\") pod \"406832ee-9a27-4930-a47e-0fc8cf4b3a86\" (UID: \"406832ee-9a27-4930-a47e-0fc8cf4b3a86\") " Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.345418 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc2pt\" (UniqueName: \"kubernetes.io/projected/cfd139ae-cd72-4d79-9fd2-ad5f199a33be-kube-api-access-dc2pt\") pod \"cfd139ae-cd72-4d79-9fd2-ad5f199a33be\" (UID: \"cfd139ae-cd72-4d79-9fd2-ad5f199a33be\") " Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.346436 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fda3961-30ca-44c7-94f0-591558331b7d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2fda3961-30ca-44c7-94f0-591558331b7d" (UID: "2fda3961-30ca-44c7-94f0-591558331b7d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.346868 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e229dd0-7f9f-4705-804d-584cbb33869c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3e229dd0-7f9f-4705-804d-584cbb33869c" (UID: "3e229dd0-7f9f-4705-804d-584cbb33869c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.347269 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfd139ae-cd72-4d79-9fd2-ad5f199a33be-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cfd139ae-cd72-4d79-9fd2-ad5f199a33be" (UID: "cfd139ae-cd72-4d79-9fd2-ad5f199a33be"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.348368 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/406832ee-9a27-4930-a47e-0fc8cf4b3a86-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "406832ee-9a27-4930-a47e-0fc8cf4b3a86" (UID: "406832ee-9a27-4930-a47e-0fc8cf4b3a86"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.349563 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21f00b80-abd7-49b1-8323-1f6d72a02455-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "21f00b80-abd7-49b1-8323-1f6d72a02455" (UID: "21f00b80-abd7-49b1-8323-1f6d72a02455"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.349836 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e229dd0-7f9f-4705-804d-584cbb33869c-kube-api-access-cd4g7" (OuterVolumeSpecName: "kube-api-access-cd4g7") pod "3e229dd0-7f9f-4705-804d-584cbb33869c" (UID: "3e229dd0-7f9f-4705-804d-584cbb33869c"). InnerVolumeSpecName "kube-api-access-cd4g7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.350952 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfd139ae-cd72-4d79-9fd2-ad5f199a33be-kube-api-access-dc2pt" (OuterVolumeSpecName: "kube-api-access-dc2pt") pod "cfd139ae-cd72-4d79-9fd2-ad5f199a33be" (UID: "cfd139ae-cd72-4d79-9fd2-ad5f199a33be"). InnerVolumeSpecName "kube-api-access-dc2pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.354360 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21f00b80-abd7-49b1-8323-1f6d72a02455-kube-api-access-l62tp" (OuterVolumeSpecName: "kube-api-access-l62tp") pod "21f00b80-abd7-49b1-8323-1f6d72a02455" (UID: "21f00b80-abd7-49b1-8323-1f6d72a02455"). InnerVolumeSpecName "kube-api-access-l62tp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.355426 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fda3961-30ca-44c7-94f0-591558331b7d-kube-api-access-t9qdd" (OuterVolumeSpecName: "kube-api-access-t9qdd") pod "2fda3961-30ca-44c7-94f0-591558331b7d" (UID: "2fda3961-30ca-44c7-94f0-591558331b7d"). InnerVolumeSpecName "kube-api-access-t9qdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.355500 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/406832ee-9a27-4930-a47e-0fc8cf4b3a86-kube-api-access-ctpnw" (OuterVolumeSpecName: "kube-api-access-ctpnw") pod "406832ee-9a27-4930-a47e-0fc8cf4b3a86" (UID: "406832ee-9a27-4930-a47e-0fc8cf4b3a86"). InnerVolumeSpecName "kube-api-access-ctpnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.448857 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9qdd\" (UniqueName: \"kubernetes.io/projected/2fda3961-30ca-44c7-94f0-591558331b7d-kube-api-access-t9qdd\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.449217 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fda3961-30ca-44c7-94f0-591558331b7d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.449301 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cfd139ae-cd72-4d79-9fd2-ad5f199a33be-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.449396 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/406832ee-9a27-4930-a47e-0fc8cf4b3a86-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.449473 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l62tp\" (UniqueName: \"kubernetes.io/projected/21f00b80-abd7-49b1-8323-1f6d72a02455-kube-api-access-l62tp\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.449550 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e229dd0-7f9f-4705-804d-584cbb33869c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.449674 4646 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21f00b80-abd7-49b1-8323-1f6d72a02455-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.449827 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd4g7\" (UniqueName: \"kubernetes.io/projected/3e229dd0-7f9f-4705-804d-584cbb33869c-kube-api-access-cd4g7\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.449953 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctpnw\" (UniqueName: \"kubernetes.io/projected/406832ee-9a27-4930-a47e-0fc8cf4b3a86-kube-api-access-ctpnw\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:21 crc kubenswrapper[4646]: I1203 11:14:21.450058 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc2pt\" (UniqueName: \"kubernetes.io/projected/cfd139ae-cd72-4d79-9fd2-ad5f199a33be-kube-api-access-dc2pt\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:22 crc kubenswrapper[4646]: I1203 11:14:22.944410 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:14:22 crc kubenswrapper[4646]: I1203 11:14:22.945055 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="ceilometer-central-agent" containerID="cri-o://9d3f3008df0625fdf188614b4f01c299dab949acd190092206d2244869a11277" gracePeriod=30 Dec 03 11:14:22 crc kubenswrapper[4646]: I1203 11:14:22.945088 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="sg-core" containerID="cri-o://123f7f2d16586d8ef519036a810613223a2a93969fcb5f188529c6296ddd255a" gracePeriod=30 Dec 03 11:14:22 crc kubenswrapper[4646]: I1203 11:14:22.945176 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="proxy-httpd" containerID="cri-o://4192479ffb79202d352ccc6c1693bdb5c0524f53d626432908e0f22fadc730f3" gracePeriod=30 Dec 03 11:14:22 crc kubenswrapper[4646]: I1203 11:14:22.945233 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="ceilometer-notification-agent" containerID="cri-o://0c97fe281a42412bdce76422505dcb9eba9427ee9e31d63d8f44a2c93fbfce1c" gracePeriod=30 Dec 03 11:14:23 crc kubenswrapper[4646]: I1203 11:14:23.355654 4646 generic.go:334] "Generic (PLEG): container finished" podID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerID="4192479ffb79202d352ccc6c1693bdb5c0524f53d626432908e0f22fadc730f3" exitCode=0 Dec 03 11:14:23 crc kubenswrapper[4646]: I1203 11:14:23.355693 4646 generic.go:334] "Generic (PLEG): container finished" podID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerID="123f7f2d16586d8ef519036a810613223a2a93969fcb5f188529c6296ddd255a" exitCode=2 Dec 03 11:14:23 crc kubenswrapper[4646]: I1203 11:14:23.355718 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16532a7f-0c9f-4a66-83b0-a3f1d731ac94","Type":"ContainerDied","Data":"4192479ffb79202d352ccc6c1693bdb5c0524f53d626432908e0f22fadc730f3"} Dec 03 11:14:23 crc kubenswrapper[4646]: I1203 11:14:23.355748 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16532a7f-0c9f-4a66-83b0-a3f1d731ac94","Type":"ContainerDied","Data":"123f7f2d16586d8ef519036a810613223a2a93969fcb5f188529c6296ddd255a"} Dec 03 11:14:24 crc kubenswrapper[4646]: I1203 11:14:24.373159 4646 generic.go:334] "Generic (PLEG): container finished" podID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerID="0c97fe281a42412bdce76422505dcb9eba9427ee9e31d63d8f44a2c93fbfce1c" exitCode=0 Dec 03 11:14:24 crc kubenswrapper[4646]: I1203 11:14:24.373234 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16532a7f-0c9f-4a66-83b0-a3f1d731ac94","Type":"ContainerDied","Data":"0c97fe281a42412bdce76422505dcb9eba9427ee9e31d63d8f44a2c93fbfce1c"} Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.193909 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.312497 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-config-data\") pod \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.312805 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-scripts\") pod \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.312943 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-combined-ca-bundle\") pod \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.313028 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-sg-core-conf-yaml\") pod \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.313175 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-run-httpd\") pod \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.313267 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-log-httpd\") pod \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.313382 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpcxs\" (UniqueName: \"kubernetes.io/projected/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-kube-api-access-vpcxs\") pod \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\" (UID: \"16532a7f-0c9f-4a66-83b0-a3f1d731ac94\") " Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.313812 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "16532a7f-0c9f-4a66-83b0-a3f1d731ac94" (UID: "16532a7f-0c9f-4a66-83b0-a3f1d731ac94"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.314038 4646 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.314145 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "16532a7f-0c9f-4a66-83b0-a3f1d731ac94" (UID: "16532a7f-0c9f-4a66-83b0-a3f1d731ac94"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.326541 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-scripts" (OuterVolumeSpecName: "scripts") pod "16532a7f-0c9f-4a66-83b0-a3f1d731ac94" (UID: "16532a7f-0c9f-4a66-83b0-a3f1d731ac94"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.326767 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-kube-api-access-vpcxs" (OuterVolumeSpecName: "kube-api-access-vpcxs") pod "16532a7f-0c9f-4a66-83b0-a3f1d731ac94" (UID: "16532a7f-0c9f-4a66-83b0-a3f1d731ac94"). InnerVolumeSpecName "kube-api-access-vpcxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.339802 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "16532a7f-0c9f-4a66-83b0-a3f1d731ac94" (UID: "16532a7f-0c9f-4a66-83b0-a3f1d731ac94"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.379323 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16532a7f-0c9f-4a66-83b0-a3f1d731ac94" (UID: "16532a7f-0c9f-4a66-83b0-a3f1d731ac94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.386692 4646 generic.go:334] "Generic (PLEG): container finished" podID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerID="9d3f3008df0625fdf188614b4f01c299dab949acd190092206d2244869a11277" exitCode=0 Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.386751 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16532a7f-0c9f-4a66-83b0-a3f1d731ac94","Type":"ContainerDied","Data":"9d3f3008df0625fdf188614b4f01c299dab949acd190092206d2244869a11277"} Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.386780 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16532a7f-0c9f-4a66-83b0-a3f1d731ac94","Type":"ContainerDied","Data":"28cc20390276c58b8adb94f7aad3a4f4d611f0bac6d444868ced3fe707401270"} Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.386795 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.386817 4646 scope.go:117] "RemoveContainer" containerID="4192479ffb79202d352ccc6c1693bdb5c0524f53d626432908e0f22fadc730f3" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.406618 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-config-data" (OuterVolumeSpecName: "config-data") pod "16532a7f-0c9f-4a66-83b0-a3f1d731ac94" (UID: "16532a7f-0c9f-4a66-83b0-a3f1d731ac94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.412050 4646 scope.go:117] "RemoveContainer" containerID="123f7f2d16586d8ef519036a810613223a2a93969fcb5f188529c6296ddd255a" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.415545 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.415665 4646 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.415726 4646 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.415779 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpcxs\" (UniqueName: \"kubernetes.io/projected/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-kube-api-access-vpcxs\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.415907 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.416005 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16532a7f-0c9f-4a66-83b0-a3f1d731ac94-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.428959 4646 scope.go:117] "RemoveContainer" containerID="0c97fe281a42412bdce76422505dcb9eba9427ee9e31d63d8f44a2c93fbfce1c" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.445553 4646 scope.go:117] "RemoveContainer" containerID="9d3f3008df0625fdf188614b4f01c299dab949acd190092206d2244869a11277" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.463046 4646 scope.go:117] "RemoveContainer" containerID="4192479ffb79202d352ccc6c1693bdb5c0524f53d626432908e0f22fadc730f3" Dec 03 11:14:25 crc kubenswrapper[4646]: E1203 11:14:25.463524 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4192479ffb79202d352ccc6c1693bdb5c0524f53d626432908e0f22fadc730f3\": container with ID starting with 4192479ffb79202d352ccc6c1693bdb5c0524f53d626432908e0f22fadc730f3 not found: ID does not exist" containerID="4192479ffb79202d352ccc6c1693bdb5c0524f53d626432908e0f22fadc730f3" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.463576 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4192479ffb79202d352ccc6c1693bdb5c0524f53d626432908e0f22fadc730f3"} err="failed to get container status \"4192479ffb79202d352ccc6c1693bdb5c0524f53d626432908e0f22fadc730f3\": rpc error: code = NotFound desc = could not find container \"4192479ffb79202d352ccc6c1693bdb5c0524f53d626432908e0f22fadc730f3\": container with ID starting with 4192479ffb79202d352ccc6c1693bdb5c0524f53d626432908e0f22fadc730f3 not found: ID does not exist" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.463609 4646 scope.go:117] "RemoveContainer" containerID="123f7f2d16586d8ef519036a810613223a2a93969fcb5f188529c6296ddd255a" Dec 03 11:14:25 crc kubenswrapper[4646]: E1203 11:14:25.464018 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"123f7f2d16586d8ef519036a810613223a2a93969fcb5f188529c6296ddd255a\": container with ID starting with 123f7f2d16586d8ef519036a810613223a2a93969fcb5f188529c6296ddd255a not found: ID does not exist" containerID="123f7f2d16586d8ef519036a810613223a2a93969fcb5f188529c6296ddd255a" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.464064 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"123f7f2d16586d8ef519036a810613223a2a93969fcb5f188529c6296ddd255a"} err="failed to get container status \"123f7f2d16586d8ef519036a810613223a2a93969fcb5f188529c6296ddd255a\": rpc error: code = NotFound desc = could not find container \"123f7f2d16586d8ef519036a810613223a2a93969fcb5f188529c6296ddd255a\": container with ID starting with 123f7f2d16586d8ef519036a810613223a2a93969fcb5f188529c6296ddd255a not found: ID does not exist" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.464096 4646 scope.go:117] "RemoveContainer" containerID="0c97fe281a42412bdce76422505dcb9eba9427ee9e31d63d8f44a2c93fbfce1c" Dec 03 11:14:25 crc kubenswrapper[4646]: E1203 11:14:25.464385 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c97fe281a42412bdce76422505dcb9eba9427ee9e31d63d8f44a2c93fbfce1c\": container with ID starting with 0c97fe281a42412bdce76422505dcb9eba9427ee9e31d63d8f44a2c93fbfce1c not found: ID does not exist" containerID="0c97fe281a42412bdce76422505dcb9eba9427ee9e31d63d8f44a2c93fbfce1c" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.464419 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c97fe281a42412bdce76422505dcb9eba9427ee9e31d63d8f44a2c93fbfce1c"} err="failed to get container status \"0c97fe281a42412bdce76422505dcb9eba9427ee9e31d63d8f44a2c93fbfce1c\": rpc error: code = NotFound desc = could not find container \"0c97fe281a42412bdce76422505dcb9eba9427ee9e31d63d8f44a2c93fbfce1c\": container with ID starting with 0c97fe281a42412bdce76422505dcb9eba9427ee9e31d63d8f44a2c93fbfce1c not found: ID does not exist" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.464567 4646 scope.go:117] "RemoveContainer" containerID="9d3f3008df0625fdf188614b4f01c299dab949acd190092206d2244869a11277" Dec 03 11:14:25 crc kubenswrapper[4646]: E1203 11:14:25.464853 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d3f3008df0625fdf188614b4f01c299dab949acd190092206d2244869a11277\": container with ID starting with 9d3f3008df0625fdf188614b4f01c299dab949acd190092206d2244869a11277 not found: ID does not exist" containerID="9d3f3008df0625fdf188614b4f01c299dab949acd190092206d2244869a11277" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.464895 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d3f3008df0625fdf188614b4f01c299dab949acd190092206d2244869a11277"} err="failed to get container status \"9d3f3008df0625fdf188614b4f01c299dab949acd190092206d2244869a11277\": rpc error: code = NotFound desc = could not find container \"9d3f3008df0625fdf188614b4f01c299dab949acd190092206d2244869a11277\": container with ID starting with 9d3f3008df0625fdf188614b4f01c299dab949acd190092206d2244869a11277 not found: ID does not exist" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.731460 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.738537 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.770028 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:14:25 crc kubenswrapper[4646]: E1203 11:14:25.770699 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfd139ae-cd72-4d79-9fd2-ad5f199a33be" containerName="mariadb-account-create-update" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.770825 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfd139ae-cd72-4d79-9fd2-ad5f199a33be" containerName="mariadb-account-create-update" Dec 03 11:14:25 crc kubenswrapper[4646]: E1203 11:14:25.770902 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="proxy-httpd" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.770963 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="proxy-httpd" Dec 03 11:14:25 crc kubenswrapper[4646]: E1203 11:14:25.771038 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="406832ee-9a27-4930-a47e-0fc8cf4b3a86" containerName="mariadb-account-create-update" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.771499 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="406832ee-9a27-4930-a47e-0fc8cf4b3a86" containerName="mariadb-account-create-update" Dec 03 11:14:25 crc kubenswrapper[4646]: E1203 11:14:25.771580 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21f00b80-abd7-49b1-8323-1f6d72a02455" containerName="mariadb-database-create" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.771658 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="21f00b80-abd7-49b1-8323-1f6d72a02455" containerName="mariadb-database-create" Dec 03 11:14:25 crc kubenswrapper[4646]: E1203 11:14:25.771732 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8adb3112-e79b-40a3-a906-65bd37489c7a" containerName="mariadb-database-create" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.774970 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="8adb3112-e79b-40a3-a906-65bd37489c7a" containerName="mariadb-database-create" Dec 03 11:14:25 crc kubenswrapper[4646]: E1203 11:14:25.775249 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fda3961-30ca-44c7-94f0-591558331b7d" containerName="mariadb-database-create" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.775304 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fda3961-30ca-44c7-94f0-591558331b7d" containerName="mariadb-database-create" Dec 03 11:14:25 crc kubenswrapper[4646]: E1203 11:14:25.775397 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e229dd0-7f9f-4705-804d-584cbb33869c" containerName="mariadb-account-create-update" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.775610 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e229dd0-7f9f-4705-804d-584cbb33869c" containerName="mariadb-account-create-update" Dec 03 11:14:25 crc kubenswrapper[4646]: E1203 11:14:25.775782 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="ceilometer-notification-agent" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.775854 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="ceilometer-notification-agent" Dec 03 11:14:25 crc kubenswrapper[4646]: E1203 11:14:25.775958 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="ceilometer-central-agent" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.776024 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="ceilometer-central-agent" Dec 03 11:14:25 crc kubenswrapper[4646]: E1203 11:14:25.776104 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="sg-core" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.776178 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="sg-core" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.776604 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="ceilometer-central-agent" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.776721 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e229dd0-7f9f-4705-804d-584cbb33869c" containerName="mariadb-account-create-update" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.776814 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="sg-core" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.776896 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="21f00b80-abd7-49b1-8323-1f6d72a02455" containerName="mariadb-database-create" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.777039 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="406832ee-9a27-4930-a47e-0fc8cf4b3a86" containerName="mariadb-account-create-update" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.777124 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="ceilometer-notification-agent" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.777820 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" containerName="proxy-httpd" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.777893 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="8adb3112-e79b-40a3-a906-65bd37489c7a" containerName="mariadb-database-create" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.777983 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fda3961-30ca-44c7-94f0-591558331b7d" containerName="mariadb-database-create" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.778054 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfd139ae-cd72-4d79-9fd2-ad5f199a33be" containerName="mariadb-account-create-update" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.780301 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.789127 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.789616 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.792230 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.860834 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16532a7f-0c9f-4a66-83b0-a3f1d731ac94" path="/var/lib/kubelet/pods/16532a7f-0c9f-4a66-83b0-a3f1d731ac94/volumes" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.925350 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-config-data\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.925421 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-run-httpd\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.925450 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-log-httpd\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.925714 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.925758 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgrhd\" (UniqueName: \"kubernetes.io/projected/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-kube-api-access-hgrhd\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.925817 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-scripts\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.925952 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.963970 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:14:25 crc kubenswrapper[4646]: I1203 11:14:25.964029 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.027487 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.027545 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-config-data\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.027578 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-run-httpd\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.027596 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-log-httpd\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.027668 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.027684 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgrhd\" (UniqueName: \"kubernetes.io/projected/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-kube-api-access-hgrhd\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.027717 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-scripts\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.028764 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-log-httpd\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.032081 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-scripts\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.032856 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-config-data\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.033915 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-run-httpd\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.034276 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.047446 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.049730 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgrhd\" (UniqueName: \"kubernetes.io/projected/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-kube-api-access-hgrhd\") pod \"ceilometer-0\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " pod="openstack/ceilometer-0" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.123911 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.543773 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:14:26 crc kubenswrapper[4646]: W1203 11:14:26.546053 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7bddeee_2832_4e2d_9cd9_0cc05f781a2c.slice/crio-ab006f93874712803df7ce50bcc8329e46d664df28287b0e9393ddf6024f81f1 WatchSource:0}: Error finding container ab006f93874712803df7ce50bcc8329e46d664df28287b0e9393ddf6024f81f1: Status 404 returned error can't find the container with id ab006f93874712803df7ce50bcc8329e46d664df28287b0e9393ddf6024f81f1 Dec 03 11:14:26 crc kubenswrapper[4646]: I1203 11:14:26.547800 4646 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.406263 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c","Type":"ContainerStarted","Data":"ab006f93874712803df7ce50bcc8329e46d664df28287b0e9393ddf6024f81f1"} Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.413224 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wnq98"] Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.414687 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.416569 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-s7jhl" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.420327 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.420559 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.439438 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wnq98"] Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.563719 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q84sp\" (UniqueName: \"kubernetes.io/projected/5eb852c6-390b-4062-8350-3d614c6e1945-kube-api-access-q84sp\") pod \"nova-cell0-conductor-db-sync-wnq98\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.563794 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wnq98\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.563847 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-config-data\") pod \"nova-cell0-conductor-db-sync-wnq98\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.564254 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-scripts\") pod \"nova-cell0-conductor-db-sync-wnq98\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.665325 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-scripts\") pod \"nova-cell0-conductor-db-sync-wnq98\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.665408 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q84sp\" (UniqueName: \"kubernetes.io/projected/5eb852c6-390b-4062-8350-3d614c6e1945-kube-api-access-q84sp\") pod \"nova-cell0-conductor-db-sync-wnq98\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.665463 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wnq98\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.665520 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-config-data\") pod \"nova-cell0-conductor-db-sync-wnq98\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.670281 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-config-data\") pod \"nova-cell0-conductor-db-sync-wnq98\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.672016 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wnq98\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.685236 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-scripts\") pod \"nova-cell0-conductor-db-sync-wnq98\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.692852 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q84sp\" (UniqueName: \"kubernetes.io/projected/5eb852c6-390b-4062-8350-3d614c6e1945-kube-api-access-q84sp\") pod \"nova-cell0-conductor-db-sync-wnq98\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:27 crc kubenswrapper[4646]: I1203 11:14:27.729476 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:28 crc kubenswrapper[4646]: I1203 11:14:28.227204 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wnq98"] Dec 03 11:14:28 crc kubenswrapper[4646]: I1203 11:14:28.415454 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wnq98" event={"ID":"5eb852c6-390b-4062-8350-3d614c6e1945","Type":"ContainerStarted","Data":"65b4abda23414288a4d592ef78ed76fd263fd5bea5f94d1f6748e57454a05246"} Dec 03 11:14:28 crc kubenswrapper[4646]: I1203 11:14:28.417788 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c","Type":"ContainerStarted","Data":"de8d9e339b358ec44d4e0dad1b2be6569b0bd7ca298d68021356ca4e5add294f"} Dec 03 11:14:28 crc kubenswrapper[4646]: I1203 11:14:28.418703 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c","Type":"ContainerStarted","Data":"750d3b08a3b61d3cd698166b661c653dc1dde85b2a143951b403d255dfecbf8a"} Dec 03 11:14:29 crc kubenswrapper[4646]: I1203 11:14:29.438573 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c","Type":"ContainerStarted","Data":"00190acf33ab3e604c60184a8c7350d5560756bab9bc79ad84634d8c99d767ff"} Dec 03 11:14:31 crc kubenswrapper[4646]: I1203 11:14:31.456512 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c","Type":"ContainerStarted","Data":"21b24b41604473c38dad81d0cffa3e5fa60589f2862a7460eb04b4992f372315"} Dec 03 11:14:31 crc kubenswrapper[4646]: I1203 11:14:31.457133 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 11:14:31 crc kubenswrapper[4646]: I1203 11:14:31.481566 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.362936581 podStartE2EDuration="6.481548103s" podCreationTimestamp="2025-12-03 11:14:25 +0000 UTC" firstStartedPulling="2025-12-03 11:14:26.547510524 +0000 UTC m=+1243.010566659" lastFinishedPulling="2025-12-03 11:14:30.666122046 +0000 UTC m=+1247.129178181" observedRunningTime="2025-12-03 11:14:31.477855679 +0000 UTC m=+1247.940911824" watchObservedRunningTime="2025-12-03 11:14:31.481548103 +0000 UTC m=+1247.944604238" Dec 03 11:14:37 crc kubenswrapper[4646]: I1203 11:14:37.510737 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wnq98" event={"ID":"5eb852c6-390b-4062-8350-3d614c6e1945","Type":"ContainerStarted","Data":"a3d8337b5dbac49a804da837d377dd6fe9de1d2460e833c3306196485aca5749"} Dec 03 11:14:37 crc kubenswrapper[4646]: I1203 11:14:37.532544 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-wnq98" podStartSLOduration=1.6426884510000002 podStartE2EDuration="10.532525933s" podCreationTimestamp="2025-12-03 11:14:27 +0000 UTC" firstStartedPulling="2025-12-03 11:14:28.23630432 +0000 UTC m=+1244.699360455" lastFinishedPulling="2025-12-03 11:14:37.126141802 +0000 UTC m=+1253.589197937" observedRunningTime="2025-12-03 11:14:37.530231079 +0000 UTC m=+1253.993287214" watchObservedRunningTime="2025-12-03 11:14:37.532525933 +0000 UTC m=+1253.995582068" Dec 03 11:14:48 crc kubenswrapper[4646]: I1203 11:14:48.601262 4646 generic.go:334] "Generic (PLEG): container finished" podID="5eb852c6-390b-4062-8350-3d614c6e1945" containerID="a3d8337b5dbac49a804da837d377dd6fe9de1d2460e833c3306196485aca5749" exitCode=0 Dec 03 11:14:48 crc kubenswrapper[4646]: I1203 11:14:48.601346 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wnq98" event={"ID":"5eb852c6-390b-4062-8350-3d614c6e1945","Type":"ContainerDied","Data":"a3d8337b5dbac49a804da837d377dd6fe9de1d2460e833c3306196485aca5749"} Dec 03 11:14:49 crc kubenswrapper[4646]: I1203 11:14:49.926603 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.080614 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-combined-ca-bundle\") pod \"5eb852c6-390b-4062-8350-3d614c6e1945\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.080783 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q84sp\" (UniqueName: \"kubernetes.io/projected/5eb852c6-390b-4062-8350-3d614c6e1945-kube-api-access-q84sp\") pod \"5eb852c6-390b-4062-8350-3d614c6e1945\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.080899 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-config-data\") pod \"5eb852c6-390b-4062-8350-3d614c6e1945\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.080924 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-scripts\") pod \"5eb852c6-390b-4062-8350-3d614c6e1945\" (UID: \"5eb852c6-390b-4062-8350-3d614c6e1945\") " Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.089178 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eb852c6-390b-4062-8350-3d614c6e1945-kube-api-access-q84sp" (OuterVolumeSpecName: "kube-api-access-q84sp") pod "5eb852c6-390b-4062-8350-3d614c6e1945" (UID: "5eb852c6-390b-4062-8350-3d614c6e1945"). InnerVolumeSpecName "kube-api-access-q84sp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.091472 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-scripts" (OuterVolumeSpecName: "scripts") pod "5eb852c6-390b-4062-8350-3d614c6e1945" (UID: "5eb852c6-390b-4062-8350-3d614c6e1945"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.114916 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-config-data" (OuterVolumeSpecName: "config-data") pod "5eb852c6-390b-4062-8350-3d614c6e1945" (UID: "5eb852c6-390b-4062-8350-3d614c6e1945"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.116281 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5eb852c6-390b-4062-8350-3d614c6e1945" (UID: "5eb852c6-390b-4062-8350-3d614c6e1945"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.182657 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.182683 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q84sp\" (UniqueName: \"kubernetes.io/projected/5eb852c6-390b-4062-8350-3d614c6e1945-kube-api-access-q84sp\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.182694 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.182702 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5eb852c6-390b-4062-8350-3d614c6e1945-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.620979 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wnq98" event={"ID":"5eb852c6-390b-4062-8350-3d614c6e1945","Type":"ContainerDied","Data":"65b4abda23414288a4d592ef78ed76fd263fd5bea5f94d1f6748e57454a05246"} Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.621305 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65b4abda23414288a4d592ef78ed76fd263fd5bea5f94d1f6748e57454a05246" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.621080 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wnq98" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.744539 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 11:14:50 crc kubenswrapper[4646]: E1203 11:14:50.744969 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eb852c6-390b-4062-8350-3d614c6e1945" containerName="nova-cell0-conductor-db-sync" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.744990 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eb852c6-390b-4062-8350-3d614c6e1945" containerName="nova-cell0-conductor-db-sync" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.745167 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eb852c6-390b-4062-8350-3d614c6e1945" containerName="nova-cell0-conductor-db-sync" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.745923 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.747588 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.750370 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-s7jhl" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.763146 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.896796 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mpsc\" (UniqueName: \"kubernetes.io/projected/c0b8e9c3-6587-4b17-a294-9d923deb1047-kube-api-access-5mpsc\") pod \"nova-cell0-conductor-0\" (UID: \"c0b8e9c3-6587-4b17-a294-9d923deb1047\") " pod="openstack/nova-cell0-conductor-0" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.896893 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0b8e9c3-6587-4b17-a294-9d923deb1047-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c0b8e9c3-6587-4b17-a294-9d923deb1047\") " pod="openstack/nova-cell0-conductor-0" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.897328 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0b8e9c3-6587-4b17-a294-9d923deb1047-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c0b8e9c3-6587-4b17-a294-9d923deb1047\") " pod="openstack/nova-cell0-conductor-0" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.998992 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mpsc\" (UniqueName: \"kubernetes.io/projected/c0b8e9c3-6587-4b17-a294-9d923deb1047-kube-api-access-5mpsc\") pod \"nova-cell0-conductor-0\" (UID: \"c0b8e9c3-6587-4b17-a294-9d923deb1047\") " pod="openstack/nova-cell0-conductor-0" Dec 03 11:14:50 crc kubenswrapper[4646]: I1203 11:14:50.999102 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0b8e9c3-6587-4b17-a294-9d923deb1047-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c0b8e9c3-6587-4b17-a294-9d923deb1047\") " pod="openstack/nova-cell0-conductor-0" Dec 03 11:14:51 crc kubenswrapper[4646]: I1203 11:14:50.999166 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0b8e9c3-6587-4b17-a294-9d923deb1047-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c0b8e9c3-6587-4b17-a294-9d923deb1047\") " pod="openstack/nova-cell0-conductor-0" Dec 03 11:14:51 crc kubenswrapper[4646]: I1203 11:14:51.003851 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0b8e9c3-6587-4b17-a294-9d923deb1047-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c0b8e9c3-6587-4b17-a294-9d923deb1047\") " pod="openstack/nova-cell0-conductor-0" Dec 03 11:14:51 crc kubenswrapper[4646]: I1203 11:14:51.012350 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0b8e9c3-6587-4b17-a294-9d923deb1047-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c0b8e9c3-6587-4b17-a294-9d923deb1047\") " pod="openstack/nova-cell0-conductor-0" Dec 03 11:14:51 crc kubenswrapper[4646]: I1203 11:14:51.015402 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mpsc\" (UniqueName: \"kubernetes.io/projected/c0b8e9c3-6587-4b17-a294-9d923deb1047-kube-api-access-5mpsc\") pod \"nova-cell0-conductor-0\" (UID: \"c0b8e9c3-6587-4b17-a294-9d923deb1047\") " pod="openstack/nova-cell0-conductor-0" Dec 03 11:14:51 crc kubenswrapper[4646]: I1203 11:14:51.076801 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 11:14:51 crc kubenswrapper[4646]: I1203 11:14:51.510249 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 11:14:51 crc kubenswrapper[4646]: I1203 11:14:51.629327 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c0b8e9c3-6587-4b17-a294-9d923deb1047","Type":"ContainerStarted","Data":"b6101d761b2a00da712b49fd0d8b61785abdd1d5f2144384657dbe25652671cd"} Dec 03 11:14:52 crc kubenswrapper[4646]: I1203 11:14:52.641369 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c0b8e9c3-6587-4b17-a294-9d923deb1047","Type":"ContainerStarted","Data":"704daba812ce2e28aabf46750c7557192388f2e2ab0ac9c500f6981ec4de0737"} Dec 03 11:14:52 crc kubenswrapper[4646]: I1203 11:14:52.641818 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 11:14:52 crc kubenswrapper[4646]: I1203 11:14:52.661275 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.661256711 podStartE2EDuration="2.661256711s" podCreationTimestamp="2025-12-03 11:14:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:14:52.657641129 +0000 UTC m=+1269.120697254" watchObservedRunningTime="2025-12-03 11:14:52.661256711 +0000 UTC m=+1269.124312846" Dec 03 11:14:55 crc kubenswrapper[4646]: I1203 11:14:55.965004 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:14:55 crc kubenswrapper[4646]: I1203 11:14:55.965106 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:14:55 crc kubenswrapper[4646]: I1203 11:14:55.965168 4646 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 11:14:55 crc kubenswrapper[4646]: I1203 11:14:55.966239 4646 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dfd3243f7f8d357efda5aacf86cad097d4462b47d7d216b50c99ea94ea34af46"} pod="openshift-machine-config-operator/machine-config-daemon-pndzt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 11:14:55 crc kubenswrapper[4646]: I1203 11:14:55.966319 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" containerID="cri-o://dfd3243f7f8d357efda5aacf86cad097d4462b47d7d216b50c99ea94ea34af46" gracePeriod=600 Dec 03 11:14:56 crc kubenswrapper[4646]: I1203 11:14:56.107994 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 11:14:56 crc kubenswrapper[4646]: I1203 11:14:56.142697 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 11:14:56 crc kubenswrapper[4646]: I1203 11:14:56.680513 4646 generic.go:334] "Generic (PLEG): container finished" podID="9695d947-fc43-4625-86af-a1a3e177bb27" containerID="dfd3243f7f8d357efda5aacf86cad097d4462b47d7d216b50c99ea94ea34af46" exitCode=0 Dec 03 11:14:56 crc kubenswrapper[4646]: I1203 11:14:56.680537 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerDied","Data":"dfd3243f7f8d357efda5aacf86cad097d4462b47d7d216b50c99ea94ea34af46"} Dec 03 11:14:56 crc kubenswrapper[4646]: I1203 11:14:56.680594 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerStarted","Data":"1e85145461d5955667430f06bb56cf842506506ca5bbed595179c9c9f545d13b"} Dec 03 11:14:56 crc kubenswrapper[4646]: I1203 11:14:56.680614 4646 scope.go:117] "RemoveContainer" containerID="14e17deb53ff7184afa03eb1fff9074fe416c1136352cde34645028a19901f16" Dec 03 11:14:56 crc kubenswrapper[4646]: I1203 11:14:56.812274 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-vx27k"] Dec 03 11:14:56 crc kubenswrapper[4646]: I1203 11:14:56.813678 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:14:56 crc kubenswrapper[4646]: I1203 11:14:56.819066 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 03 11:14:56 crc kubenswrapper[4646]: I1203 11:14:56.819320 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 03 11:14:56 crc kubenswrapper[4646]: I1203 11:14:56.837901 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vx27k"] Dec 03 11:14:56 crc kubenswrapper[4646]: I1203 11:14:56.903212 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vx27k\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:14:56 crc kubenswrapper[4646]: I1203 11:14:56.903253 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrlnz\" (UniqueName: \"kubernetes.io/projected/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-kube-api-access-vrlnz\") pod \"nova-cell0-cell-mapping-vx27k\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:14:56 crc kubenswrapper[4646]: I1203 11:14:56.903438 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-config-data\") pod \"nova-cell0-cell-mapping-vx27k\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:14:56 crc kubenswrapper[4646]: I1203 11:14:56.903499 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-scripts\") pod \"nova-cell0-cell-mapping-vx27k\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.004574 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vx27k\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.004893 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrlnz\" (UniqueName: \"kubernetes.io/projected/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-kube-api-access-vrlnz\") pod \"nova-cell0-cell-mapping-vx27k\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.004990 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-config-data\") pod \"nova-cell0-cell-mapping-vx27k\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.005028 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-scripts\") pod \"nova-cell0-cell-mapping-vx27k\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.014818 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vx27k\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.014946 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-config-data\") pod \"nova-cell0-cell-mapping-vx27k\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.015274 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-scripts\") pod \"nova-cell0-cell-mapping-vx27k\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.035070 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.036189 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.047865 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.090981 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrlnz\" (UniqueName: \"kubernetes.io/projected/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-kube-api-access-vrlnz\") pod \"nova-cell0-cell-mapping-vx27k\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.106189 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46aeaecf-b036-48cb-8b33-04c134aa5b8b-config-data\") pod \"nova-scheduler-0\" (UID: \"46aeaecf-b036-48cb-8b33-04c134aa5b8b\") " pod="openstack/nova-scheduler-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.106371 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6dtp\" (UniqueName: \"kubernetes.io/projected/46aeaecf-b036-48cb-8b33-04c134aa5b8b-kube-api-access-c6dtp\") pod \"nova-scheduler-0\" (UID: \"46aeaecf-b036-48cb-8b33-04c134aa5b8b\") " pod="openstack/nova-scheduler-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.106438 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46aeaecf-b036-48cb-8b33-04c134aa5b8b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"46aeaecf-b036-48cb-8b33-04c134aa5b8b\") " pod="openstack/nova-scheduler-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.143416 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.145249 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.145268 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.155653 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.168787 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.186796 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.209601 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6dtp\" (UniqueName: \"kubernetes.io/projected/46aeaecf-b036-48cb-8b33-04c134aa5b8b-kube-api-access-c6dtp\") pod \"nova-scheduler-0\" (UID: \"46aeaecf-b036-48cb-8b33-04c134aa5b8b\") " pod="openstack/nova-scheduler-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.209804 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5ea85e-a978-410c-90ed-2d1f95c2083c-logs\") pod \"nova-api-0\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " pod="openstack/nova-api-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.209930 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46aeaecf-b036-48cb-8b33-04c134aa5b8b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"46aeaecf-b036-48cb-8b33-04c134aa5b8b\") " pod="openstack/nova-scheduler-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.209964 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5ea85e-a978-410c-90ed-2d1f95c2083c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " pod="openstack/nova-api-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.210039 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46aeaecf-b036-48cb-8b33-04c134aa5b8b-config-data\") pod \"nova-scheduler-0\" (UID: \"46aeaecf-b036-48cb-8b33-04c134aa5b8b\") " pod="openstack/nova-scheduler-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.210128 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckhjj\" (UniqueName: \"kubernetes.io/projected/6a5ea85e-a978-410c-90ed-2d1f95c2083c-kube-api-access-ckhjj\") pod \"nova-api-0\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " pod="openstack/nova-api-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.210199 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5ea85e-a978-410c-90ed-2d1f95c2083c-config-data\") pod \"nova-api-0\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " pod="openstack/nova-api-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.218224 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46aeaecf-b036-48cb-8b33-04c134aa5b8b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"46aeaecf-b036-48cb-8b33-04c134aa5b8b\") " pod="openstack/nova-scheduler-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.241893 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46aeaecf-b036-48cb-8b33-04c134aa5b8b-config-data\") pod \"nova-scheduler-0\" (UID: \"46aeaecf-b036-48cb-8b33-04c134aa5b8b\") " pod="openstack/nova-scheduler-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.268092 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6dtp\" (UniqueName: \"kubernetes.io/projected/46aeaecf-b036-48cb-8b33-04c134aa5b8b-kube-api-access-c6dtp\") pod \"nova-scheduler-0\" (UID: \"46aeaecf-b036-48cb-8b33-04c134aa5b8b\") " pod="openstack/nova-scheduler-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.295124 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.296569 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.309667 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.311433 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5ea85e-a978-410c-90ed-2d1f95c2083c-logs\") pod \"nova-api-0\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " pod="openstack/nova-api-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.311497 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5ea85e-a978-410c-90ed-2d1f95c2083c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " pod="openstack/nova-api-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.311549 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckhjj\" (UniqueName: \"kubernetes.io/projected/6a5ea85e-a978-410c-90ed-2d1f95c2083c-kube-api-access-ckhjj\") pod \"nova-api-0\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " pod="openstack/nova-api-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.311579 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5ea85e-a978-410c-90ed-2d1f95c2083c-config-data\") pod \"nova-api-0\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " pod="openstack/nova-api-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.319572 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5ea85e-a978-410c-90ed-2d1f95c2083c-config-data\") pod \"nova-api-0\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " pod="openstack/nova-api-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.320271 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5ea85e-a978-410c-90ed-2d1f95c2083c-logs\") pod \"nova-api-0\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " pod="openstack/nova-api-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.343910 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.344859 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5ea85e-a978-410c-90ed-2d1f95c2083c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " pod="openstack/nova-api-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.379287 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckhjj\" (UniqueName: \"kubernetes.io/projected/6a5ea85e-a978-410c-90ed-2d1f95c2083c-kube-api-access-ckhjj\") pod \"nova-api-0\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " pod="openstack/nova-api-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.416174 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp7qc\" (UniqueName: \"kubernetes.io/projected/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-kube-api-access-dp7qc\") pod \"nova-metadata-0\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " pod="openstack/nova-metadata-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.416273 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " pod="openstack/nova-metadata-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.416302 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-logs\") pod \"nova-metadata-0\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " pod="openstack/nova-metadata-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.416470 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-config-data\") pod \"nova-metadata-0\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " pod="openstack/nova-metadata-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.443577 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.444937 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.448835 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.463466 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.463574 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.482204 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.523228 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwvk8\" (UniqueName: \"kubernetes.io/projected/97a4238d-da32-427d-91cc-16d2df65221f-kube-api-access-wwvk8\") pod \"nova-cell1-novncproxy-0\" (UID: \"97a4238d-da32-427d-91cc-16d2df65221f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.523314 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp7qc\" (UniqueName: \"kubernetes.io/projected/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-kube-api-access-dp7qc\") pod \"nova-metadata-0\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " pod="openstack/nova-metadata-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.523373 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a4238d-da32-427d-91cc-16d2df65221f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"97a4238d-da32-427d-91cc-16d2df65221f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.523554 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " pod="openstack/nova-metadata-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.523605 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-logs\") pod \"nova-metadata-0\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " pod="openstack/nova-metadata-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.523923 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a4238d-da32-427d-91cc-16d2df65221f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"97a4238d-da32-427d-91cc-16d2df65221f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.523984 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-config-data\") pod \"nova-metadata-0\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " pod="openstack/nova-metadata-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.526442 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-logs\") pod \"nova-metadata-0\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " pod="openstack/nova-metadata-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.542097 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-config-data\") pod \"nova-metadata-0\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " pod="openstack/nova-metadata-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.582432 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp7qc\" (UniqueName: \"kubernetes.io/projected/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-kube-api-access-dp7qc\") pod \"nova-metadata-0\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " pod="openstack/nova-metadata-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.605749 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " pod="openstack/nova-metadata-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.614934 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-d6p7n"] Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.617720 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.627157 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a4238d-da32-427d-91cc-16d2df65221f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"97a4238d-da32-427d-91cc-16d2df65221f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.627277 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwvk8\" (UniqueName: \"kubernetes.io/projected/97a4238d-da32-427d-91cc-16d2df65221f-kube-api-access-wwvk8\") pod \"nova-cell1-novncproxy-0\" (UID: \"97a4238d-da32-427d-91cc-16d2df65221f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.627346 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a4238d-da32-427d-91cc-16d2df65221f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"97a4238d-da32-427d-91cc-16d2df65221f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.631942 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-d6p7n"] Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.637266 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a4238d-da32-427d-91cc-16d2df65221f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"97a4238d-da32-427d-91cc-16d2df65221f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.651200 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a4238d-da32-427d-91cc-16d2df65221f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"97a4238d-da32-427d-91cc-16d2df65221f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.665036 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwvk8\" (UniqueName: \"kubernetes.io/projected/97a4238d-da32-427d-91cc-16d2df65221f-kube-api-access-wwvk8\") pod \"nova-cell1-novncproxy-0\" (UID: \"97a4238d-da32-427d-91cc-16d2df65221f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.741859 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-ovsdbserver-nb\") pod \"dnsmasq-dns-8b8cf6657-d6p7n\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.741947 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-dns-svc\") pod \"dnsmasq-dns-8b8cf6657-d6p7n\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.742017 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-ovsdbserver-sb\") pod \"dnsmasq-dns-8b8cf6657-d6p7n\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.742137 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbdvt\" (UniqueName: \"kubernetes.io/projected/16b0f468-ba7c-4c95-bac0-9a19912975c2-kube-api-access-zbdvt\") pod \"dnsmasq-dns-8b8cf6657-d6p7n\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.742220 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-config\") pod \"dnsmasq-dns-8b8cf6657-d6p7n\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.843552 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-ovsdbserver-nb\") pod \"dnsmasq-dns-8b8cf6657-d6p7n\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.843605 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-dns-svc\") pod \"dnsmasq-dns-8b8cf6657-d6p7n\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.843643 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-ovsdbserver-sb\") pod \"dnsmasq-dns-8b8cf6657-d6p7n\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.843704 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbdvt\" (UniqueName: \"kubernetes.io/projected/16b0f468-ba7c-4c95-bac0-9a19912975c2-kube-api-access-zbdvt\") pod \"dnsmasq-dns-8b8cf6657-d6p7n\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.843747 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-config\") pod \"dnsmasq-dns-8b8cf6657-d6p7n\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.845239 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-ovsdbserver-nb\") pod \"dnsmasq-dns-8b8cf6657-d6p7n\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.845481 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.849309 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-config\") pod \"dnsmasq-dns-8b8cf6657-d6p7n\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.849716 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-ovsdbserver-sb\") pod \"dnsmasq-dns-8b8cf6657-d6p7n\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.850647 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-dns-svc\") pod \"dnsmasq-dns-8b8cf6657-d6p7n\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.865929 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.871598 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbdvt\" (UniqueName: \"kubernetes.io/projected/16b0f468-ba7c-4c95-bac0-9a19912975c2-kube-api-access-zbdvt\") pod \"dnsmasq-dns-8b8cf6657-d6p7n\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:57 crc kubenswrapper[4646]: I1203 11:14:57.969759 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.009471 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vx27k"] Dec 03 11:14:58 crc kubenswrapper[4646]: W1203 11:14:58.028486 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf37ef5fc_6976_46cf_87ee_6c3d8acf75a8.slice/crio-0f65f2400fcd031a401f817e1d15bc59e590cff42aeefd6bc2ef5e3e3144c406 WatchSource:0}: Error finding container 0f65f2400fcd031a401f817e1d15bc59e590cff42aeefd6bc2ef5e3e3144c406: Status 404 returned error can't find the container with id 0f65f2400fcd031a401f817e1d15bc59e590cff42aeefd6bc2ef5e3e3144c406 Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.082506 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 11:14:58 crc kubenswrapper[4646]: W1203 11:14:58.110733 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46aeaecf_b036_48cb_8b33_04c134aa5b8b.slice/crio-d7c232213893431e48fc76e9e4afbdac1bd54f910a204fd232cf0e0cacda89c5 WatchSource:0}: Error finding container d7c232213893431e48fc76e9e4afbdac1bd54f910a204fd232cf0e0cacda89c5: Status 404 returned error can't find the container with id d7c232213893431e48fc76e9e4afbdac1bd54f910a204fd232cf0e0cacda89c5 Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.188400 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:14:58 crc kubenswrapper[4646]: W1203 11:14:58.215274 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a5ea85e_a978_410c_90ed_2d1f95c2083c.slice/crio-f3909bbd4d68645b10bccf95a049c2ef45966b55f54913ca0c44985b50f43c2b WatchSource:0}: Error finding container f3909bbd4d68645b10bccf95a049c2ef45966b55f54913ca0c44985b50f43c2b: Status 404 returned error can't find the container with id f3909bbd4d68645b10bccf95a049c2ef45966b55f54913ca0c44985b50f43c2b Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.324713 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-j4gb5"] Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.325838 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.336388 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-j4gb5"] Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.342092 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.345832 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.379034 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-j4gb5\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.383738 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-config-data\") pod \"nova-cell1-conductor-db-sync-j4gb5\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.384052 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-scripts\") pod \"nova-cell1-conductor-db-sync-j4gb5\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.384287 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twl52\" (UniqueName: \"kubernetes.io/projected/56efccd7-660f-4052-afa3-dab6a63e93a2-kube-api-access-twl52\") pod \"nova-cell1-conductor-db-sync-j4gb5\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.486424 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-j4gb5\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.486587 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-config-data\") pod \"nova-cell1-conductor-db-sync-j4gb5\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.486852 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-scripts\") pod \"nova-cell1-conductor-db-sync-j4gb5\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.488261 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twl52\" (UniqueName: \"kubernetes.io/projected/56efccd7-660f-4052-afa3-dab6a63e93a2-kube-api-access-twl52\") pod \"nova-cell1-conductor-db-sync-j4gb5\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.494829 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-scripts\") pod \"nova-cell1-conductor-db-sync-j4gb5\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.494884 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-config-data\") pod \"nova-cell1-conductor-db-sync-j4gb5\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.506820 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-j4gb5\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.510134 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twl52\" (UniqueName: \"kubernetes.io/projected/56efccd7-660f-4052-afa3-dab6a63e93a2-kube-api-access-twl52\") pod \"nova-cell1-conductor-db-sync-j4gb5\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.645916 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-d6p7n"] Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.734858 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.742658 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.800508 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" event={"ID":"16b0f468-ba7c-4c95-bac0-9a19912975c2","Type":"ContainerStarted","Data":"4f715dd0c7f30e0f5688f620d8e81fcd939543930eef95db9c5436c7038f3c50"} Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.802995 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d","Type":"ContainerStarted","Data":"375cb185f3f881c959bef2486f122960b66bf539dd7325cc778d111e3faadade"} Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.804140 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6a5ea85e-a978-410c-90ed-2d1f95c2083c","Type":"ContainerStarted","Data":"f3909bbd4d68645b10bccf95a049c2ef45966b55f54913ca0c44985b50f43c2b"} Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.808950 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"46aeaecf-b036-48cb-8b33-04c134aa5b8b","Type":"ContainerStarted","Data":"d7c232213893431e48fc76e9e4afbdac1bd54f910a204fd232cf0e0cacda89c5"} Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.811514 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vx27k" event={"ID":"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8","Type":"ContainerStarted","Data":"59ab2024684ff6d3c8ff7ab17b5b5e02ba1aef3d3411b4759799dca1d9e408b4"} Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.811546 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vx27k" event={"ID":"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8","Type":"ContainerStarted","Data":"0f65f2400fcd031a401f817e1d15bc59e590cff42aeefd6bc2ef5e3e3144c406"} Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.900030 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-vx27k" podStartSLOduration=2.900013002 podStartE2EDuration="2.900013002s" podCreationTimestamp="2025-12-03 11:14:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:14:58.827563281 +0000 UTC m=+1275.290619416" watchObservedRunningTime="2025-12-03 11:14:58.900013002 +0000 UTC m=+1275.363069137" Dec 03 11:14:58 crc kubenswrapper[4646]: I1203 11:14:58.904093 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 11:14:58 crc kubenswrapper[4646]: W1203 11:14:58.907825 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97a4238d_da32_427d_91cc_16d2df65221f.slice/crio-5b94909c2e3206d64d2ce2df3f8bc35db5abbf83975362b4a8c0fe96e9362f01 WatchSource:0}: Error finding container 5b94909c2e3206d64d2ce2df3f8bc35db5abbf83975362b4a8c0fe96e9362f01: Status 404 returned error can't find the container with id 5b94909c2e3206d64d2ce2df3f8bc35db5abbf83975362b4a8c0fe96e9362f01 Dec 03 11:14:59 crc kubenswrapper[4646]: I1203 11:14:59.376780 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-j4gb5"] Dec 03 11:14:59 crc kubenswrapper[4646]: W1203 11:14:59.463983 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56efccd7_660f_4052_afa3_dab6a63e93a2.slice/crio-7c17282edadde051def412cb1c96c070106ee7946d7cfe385174590435a13797 WatchSource:0}: Error finding container 7c17282edadde051def412cb1c96c070106ee7946d7cfe385174590435a13797: Status 404 returned error can't find the container with id 7c17282edadde051def412cb1c96c070106ee7946d7cfe385174590435a13797 Dec 03 11:14:59 crc kubenswrapper[4646]: I1203 11:14:59.860928 4646 generic.go:334] "Generic (PLEG): container finished" podID="16b0f468-ba7c-4c95-bac0-9a19912975c2" containerID="32660307339de85a58f3445c2402560e1088d71134c3e721fcf99c7b1db099b0" exitCode=0 Dec 03 11:14:59 crc kubenswrapper[4646]: I1203 11:14:59.866072 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" event={"ID":"16b0f468-ba7c-4c95-bac0-9a19912975c2","Type":"ContainerDied","Data":"32660307339de85a58f3445c2402560e1088d71134c3e721fcf99c7b1db099b0"} Dec 03 11:14:59 crc kubenswrapper[4646]: I1203 11:14:59.870306 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-j4gb5" event={"ID":"56efccd7-660f-4052-afa3-dab6a63e93a2","Type":"ContainerStarted","Data":"55f494c40970f92c45b900e8e523f7be3167fa07ff2296242dd792044a06cf32"} Dec 03 11:14:59 crc kubenswrapper[4646]: I1203 11:14:59.870419 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-j4gb5" event={"ID":"56efccd7-660f-4052-afa3-dab6a63e93a2","Type":"ContainerStarted","Data":"7c17282edadde051def412cb1c96c070106ee7946d7cfe385174590435a13797"} Dec 03 11:14:59 crc kubenswrapper[4646]: I1203 11:14:59.880479 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"97a4238d-da32-427d-91cc-16d2df65221f","Type":"ContainerStarted","Data":"5b94909c2e3206d64d2ce2df3f8bc35db5abbf83975362b4a8c0fe96e9362f01"} Dec 03 11:14:59 crc kubenswrapper[4646]: I1203 11:14:59.920191 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-j4gb5" podStartSLOduration=1.9201665879999998 podStartE2EDuration="1.920166588s" podCreationTimestamp="2025-12-03 11:14:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:14:59.909307492 +0000 UTC m=+1276.372363637" watchObservedRunningTime="2025-12-03 11:14:59.920166588 +0000 UTC m=+1276.383222723" Dec 03 11:15:00 crc kubenswrapper[4646]: I1203 11:15:00.152875 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54"] Dec 03 11:15:00 crc kubenswrapper[4646]: I1203 11:15:00.154457 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" Dec 03 11:15:00 crc kubenswrapper[4646]: I1203 11:15:00.157719 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 11:15:00 crc kubenswrapper[4646]: I1203 11:15:00.158230 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 11:15:00 crc kubenswrapper[4646]: I1203 11:15:00.174731 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54"] Dec 03 11:15:00 crc kubenswrapper[4646]: I1203 11:15:00.249180 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52d1a48c-5c33-4ac1-95d6-709377ecbf30-config-volume\") pod \"collect-profiles-29412675-66h54\" (UID: \"52d1a48c-5c33-4ac1-95d6-709377ecbf30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" Dec 03 11:15:00 crc kubenswrapper[4646]: I1203 11:15:00.249253 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4hhs\" (UniqueName: \"kubernetes.io/projected/52d1a48c-5c33-4ac1-95d6-709377ecbf30-kube-api-access-s4hhs\") pod \"collect-profiles-29412675-66h54\" (UID: \"52d1a48c-5c33-4ac1-95d6-709377ecbf30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" Dec 03 11:15:00 crc kubenswrapper[4646]: I1203 11:15:00.249292 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52d1a48c-5c33-4ac1-95d6-709377ecbf30-secret-volume\") pod \"collect-profiles-29412675-66h54\" (UID: \"52d1a48c-5c33-4ac1-95d6-709377ecbf30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" Dec 03 11:15:00 crc kubenswrapper[4646]: I1203 11:15:00.352444 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52d1a48c-5c33-4ac1-95d6-709377ecbf30-config-volume\") pod \"collect-profiles-29412675-66h54\" (UID: \"52d1a48c-5c33-4ac1-95d6-709377ecbf30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" Dec 03 11:15:00 crc kubenswrapper[4646]: I1203 11:15:00.352517 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4hhs\" (UniqueName: \"kubernetes.io/projected/52d1a48c-5c33-4ac1-95d6-709377ecbf30-kube-api-access-s4hhs\") pod \"collect-profiles-29412675-66h54\" (UID: \"52d1a48c-5c33-4ac1-95d6-709377ecbf30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" Dec 03 11:15:00 crc kubenswrapper[4646]: I1203 11:15:00.352559 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52d1a48c-5c33-4ac1-95d6-709377ecbf30-secret-volume\") pod \"collect-profiles-29412675-66h54\" (UID: \"52d1a48c-5c33-4ac1-95d6-709377ecbf30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" Dec 03 11:15:00 crc kubenswrapper[4646]: I1203 11:15:00.355409 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52d1a48c-5c33-4ac1-95d6-709377ecbf30-config-volume\") pod \"collect-profiles-29412675-66h54\" (UID: \"52d1a48c-5c33-4ac1-95d6-709377ecbf30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" Dec 03 11:15:00 crc kubenswrapper[4646]: I1203 11:15:00.366205 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52d1a48c-5c33-4ac1-95d6-709377ecbf30-secret-volume\") pod \"collect-profiles-29412675-66h54\" (UID: \"52d1a48c-5c33-4ac1-95d6-709377ecbf30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" Dec 03 11:15:00 crc kubenswrapper[4646]: I1203 11:15:00.371173 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4hhs\" (UniqueName: \"kubernetes.io/projected/52d1a48c-5c33-4ac1-95d6-709377ecbf30-kube-api-access-s4hhs\") pod \"collect-profiles-29412675-66h54\" (UID: \"52d1a48c-5c33-4ac1-95d6-709377ecbf30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" Dec 03 11:15:00 crc kubenswrapper[4646]: I1203 11:15:00.505502 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" Dec 03 11:15:01 crc kubenswrapper[4646]: I1203 11:15:01.037271 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 11:15:01 crc kubenswrapper[4646]: I1203 11:15:01.037777 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="86ef1c10-7d50-4089-a463-676380821fe3" containerName="kube-state-metrics" containerID="cri-o://00ca8229142fe32f0ae234fe0b2c23dc838a16d750b14c9ead9fef9b263cdb49" gracePeriod=30 Dec 03 11:15:01 crc kubenswrapper[4646]: I1203 11:15:01.075287 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:01 crc kubenswrapper[4646]: I1203 11:15:01.091469 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 11:15:01 crc kubenswrapper[4646]: I1203 11:15:01.911637 4646 generic.go:334] "Generic (PLEG): container finished" podID="86ef1c10-7d50-4089-a463-676380821fe3" containerID="00ca8229142fe32f0ae234fe0b2c23dc838a16d750b14c9ead9fef9b263cdb49" exitCode=2 Dec 03 11:15:01 crc kubenswrapper[4646]: I1203 11:15:01.911720 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"86ef1c10-7d50-4089-a463-676380821fe3","Type":"ContainerDied","Data":"00ca8229142fe32f0ae234fe0b2c23dc838a16d750b14c9ead9fef9b263cdb49"} Dec 03 11:15:02 crc kubenswrapper[4646]: I1203 11:15:02.473522 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:15:02 crc kubenswrapper[4646]: I1203 11:15:02.473860 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="ceilometer-central-agent" containerID="cri-o://750d3b08a3b61d3cd698166b661c653dc1dde85b2a143951b403d255dfecbf8a" gracePeriod=30 Dec 03 11:15:02 crc kubenswrapper[4646]: I1203 11:15:02.473936 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="proxy-httpd" containerID="cri-o://21b24b41604473c38dad81d0cffa3e5fa60589f2862a7460eb04b4992f372315" gracePeriod=30 Dec 03 11:15:02 crc kubenswrapper[4646]: I1203 11:15:02.473974 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="ceilometer-notification-agent" containerID="cri-o://de8d9e339b358ec44d4e0dad1b2be6569b0bd7ca298d68021356ca4e5add294f" gracePeriod=30 Dec 03 11:15:02 crc kubenswrapper[4646]: I1203 11:15:02.473991 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="sg-core" containerID="cri-o://00190acf33ab3e604c60184a8c7350d5560756bab9bc79ad84634d8c99d767ff" gracePeriod=30 Dec 03 11:15:02 crc kubenswrapper[4646]: I1203 11:15:02.923209 4646 generic.go:334] "Generic (PLEG): container finished" podID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerID="00190acf33ab3e604c60184a8c7350d5560756bab9bc79ad84634d8c99d767ff" exitCode=2 Dec 03 11:15:02 crc kubenswrapper[4646]: I1203 11:15:02.923289 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c","Type":"ContainerDied","Data":"00190acf33ab3e604c60184a8c7350d5560756bab9bc79ad84634d8c99d767ff"} Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.433262 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.520750 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqgdl\" (UniqueName: \"kubernetes.io/projected/86ef1c10-7d50-4089-a463-676380821fe3-kube-api-access-nqgdl\") pod \"86ef1c10-7d50-4089-a463-676380821fe3\" (UID: \"86ef1c10-7d50-4089-a463-676380821fe3\") " Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.531694 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86ef1c10-7d50-4089-a463-676380821fe3-kube-api-access-nqgdl" (OuterVolumeSpecName: "kube-api-access-nqgdl") pod "86ef1c10-7d50-4089-a463-676380821fe3" (UID: "86ef1c10-7d50-4089-a463-676380821fe3"). InnerVolumeSpecName "kube-api-access-nqgdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.622321 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqgdl\" (UniqueName: \"kubernetes.io/projected/86ef1c10-7d50-4089-a463-676380821fe3-kube-api-access-nqgdl\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.913523 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54"] Dec 03 11:15:03 crc kubenswrapper[4646]: W1203 11:15:03.923142 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52d1a48c_5c33_4ac1_95d6_709377ecbf30.slice/crio-41f61111f5d3e15a2a3110ed73910349c93276e942081a6bd2f471355e1ddf56 WatchSource:0}: Error finding container 41f61111f5d3e15a2a3110ed73910349c93276e942081a6bd2f471355e1ddf56: Status 404 returned error can't find the container with id 41f61111f5d3e15a2a3110ed73910349c93276e942081a6bd2f471355e1ddf56 Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.947686 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"46aeaecf-b036-48cb-8b33-04c134aa5b8b","Type":"ContainerStarted","Data":"37ffdf894aa4e90c1ca1185efe46685416dfa0b2dab568f761aca9d242c6e5f2"} Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.954912 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" event={"ID":"16b0f468-ba7c-4c95-bac0-9a19912975c2","Type":"ContainerStarted","Data":"7f221a9351df0f408a440796ef0676ff07011535f8fd18dd0ced5ee80ec5b867"} Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.955462 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.956421 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d","Type":"ContainerStarted","Data":"6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed"} Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.963798 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6a5ea85e-a978-410c-90ed-2d1f95c2083c","Type":"ContainerStarted","Data":"e235ce8e8d49e6703d27bca078986c6c7514030f02ade612e8e7a1d92de748ee"} Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.975892 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"97a4238d-da32-427d-91cc-16d2df65221f","Type":"ContainerStarted","Data":"8397ac3883bdb4300287cf24c3dc1bee449b192f8df04dfee3d59fabfb9ad03c"} Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.976010 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="97a4238d-da32-427d-91cc-16d2df65221f" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://8397ac3883bdb4300287cf24c3dc1bee449b192f8df04dfee3d59fabfb9ad03c" gracePeriod=30 Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.982680 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" event={"ID":"52d1a48c-5c33-4ac1-95d6-709377ecbf30","Type":"ContainerStarted","Data":"41f61111f5d3e15a2a3110ed73910349c93276e942081a6bd2f471355e1ddf56"} Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.983890 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"86ef1c10-7d50-4089-a463-676380821fe3","Type":"ContainerDied","Data":"72c6447595701e96f7e326fffe009010a24f51e5bd989ca3ec26b2fc10b01a01"} Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.983918 4646 scope.go:117] "RemoveContainer" containerID="00ca8229142fe32f0ae234fe0b2c23dc838a16d750b14c9ead9fef9b263cdb49" Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.984013 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.988261 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.803626478 podStartE2EDuration="7.988251086s" podCreationTimestamp="2025-12-03 11:14:56 +0000 UTC" firstStartedPulling="2025-12-03 11:14:58.123491373 +0000 UTC m=+1274.586547508" lastFinishedPulling="2025-12-03 11:15:03.308115971 +0000 UTC m=+1279.771172116" observedRunningTime="2025-12-03 11:15:03.976459083 +0000 UTC m=+1280.439515218" watchObservedRunningTime="2025-12-03 11:15:03.988251086 +0000 UTC m=+1280.451307221" Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.993504 4646 generic.go:334] "Generic (PLEG): container finished" podID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerID="21b24b41604473c38dad81d0cffa3e5fa60589f2862a7460eb04b4992f372315" exitCode=0 Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.993526 4646 generic.go:334] "Generic (PLEG): container finished" podID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerID="750d3b08a3b61d3cd698166b661c653dc1dde85b2a143951b403d255dfecbf8a" exitCode=0 Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.993545 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c","Type":"ContainerDied","Data":"21b24b41604473c38dad81d0cffa3e5fa60589f2862a7460eb04b4992f372315"} Dec 03 11:15:03 crc kubenswrapper[4646]: I1203 11:15:03.993565 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c","Type":"ContainerDied","Data":"750d3b08a3b61d3cd698166b661c653dc1dde85b2a143951b403d255dfecbf8a"} Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.001549 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.5662346449999998 podStartE2EDuration="7.0015389s" podCreationTimestamp="2025-12-03 11:14:57 +0000 UTC" firstStartedPulling="2025-12-03 11:14:58.91412549 +0000 UTC m=+1275.377181625" lastFinishedPulling="2025-12-03 11:15:03.349429745 +0000 UTC m=+1279.812485880" observedRunningTime="2025-12-03 11:15:03.991268111 +0000 UTC m=+1280.454324246" watchObservedRunningTime="2025-12-03 11:15:04.0015389 +0000 UTC m=+1280.464595035" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.029175 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" podStartSLOduration=7.029159108 podStartE2EDuration="7.029159108s" podCreationTimestamp="2025-12-03 11:14:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:15:04.021519723 +0000 UTC m=+1280.484575858" watchObservedRunningTime="2025-12-03 11:15:04.029159108 +0000 UTC m=+1280.492215243" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.049588 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.062383 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.079366 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 11:15:04 crc kubenswrapper[4646]: E1203 11:15:04.079776 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ef1c10-7d50-4089-a463-676380821fe3" containerName="kube-state-metrics" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.079793 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ef1c10-7d50-4089-a463-676380821fe3" containerName="kube-state-metrics" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.079968 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="86ef1c10-7d50-4089-a463-676380821fe3" containerName="kube-state-metrics" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.080555 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.096863 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.099951 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.122217 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.133088 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb37f32e-cad3-4184-ad71-fd5cef58ca1f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"eb37f32e-cad3-4184-ad71-fd5cef58ca1f\") " pod="openstack/kube-state-metrics-0" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.133135 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/eb37f32e-cad3-4184-ad71-fd5cef58ca1f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"eb37f32e-cad3-4184-ad71-fd5cef58ca1f\") " pod="openstack/kube-state-metrics-0" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.133167 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb37f32e-cad3-4184-ad71-fd5cef58ca1f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"eb37f32e-cad3-4184-ad71-fd5cef58ca1f\") " pod="openstack/kube-state-metrics-0" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.133217 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8x27\" (UniqueName: \"kubernetes.io/projected/eb37f32e-cad3-4184-ad71-fd5cef58ca1f-kube-api-access-s8x27\") pod \"kube-state-metrics-0\" (UID: \"eb37f32e-cad3-4184-ad71-fd5cef58ca1f\") " pod="openstack/kube-state-metrics-0" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.234879 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb37f32e-cad3-4184-ad71-fd5cef58ca1f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"eb37f32e-cad3-4184-ad71-fd5cef58ca1f\") " pod="openstack/kube-state-metrics-0" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.234928 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/eb37f32e-cad3-4184-ad71-fd5cef58ca1f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"eb37f32e-cad3-4184-ad71-fd5cef58ca1f\") " pod="openstack/kube-state-metrics-0" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.234967 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb37f32e-cad3-4184-ad71-fd5cef58ca1f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"eb37f32e-cad3-4184-ad71-fd5cef58ca1f\") " pod="openstack/kube-state-metrics-0" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.235017 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8x27\" (UniqueName: \"kubernetes.io/projected/eb37f32e-cad3-4184-ad71-fd5cef58ca1f-kube-api-access-s8x27\") pod \"kube-state-metrics-0\" (UID: \"eb37f32e-cad3-4184-ad71-fd5cef58ca1f\") " pod="openstack/kube-state-metrics-0" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.240085 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb37f32e-cad3-4184-ad71-fd5cef58ca1f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"eb37f32e-cad3-4184-ad71-fd5cef58ca1f\") " pod="openstack/kube-state-metrics-0" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.240719 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/eb37f32e-cad3-4184-ad71-fd5cef58ca1f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"eb37f32e-cad3-4184-ad71-fd5cef58ca1f\") " pod="openstack/kube-state-metrics-0" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.242241 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb37f32e-cad3-4184-ad71-fd5cef58ca1f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"eb37f32e-cad3-4184-ad71-fd5cef58ca1f\") " pod="openstack/kube-state-metrics-0" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.269484 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8x27\" (UniqueName: \"kubernetes.io/projected/eb37f32e-cad3-4184-ad71-fd5cef58ca1f-kube-api-access-s8x27\") pod \"kube-state-metrics-0\" (UID: \"eb37f32e-cad3-4184-ad71-fd5cef58ca1f\") " pod="openstack/kube-state-metrics-0" Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.421214 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 11:15:04 crc kubenswrapper[4646]: W1203 11:15:04.757883 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb37f32e_cad3_4184_ad71_fd5cef58ca1f.slice/crio-26d4527404237d55c3df5c1180db71c23412fff629b2a06e5aaee4a2024dd2ff WatchSource:0}: Error finding container 26d4527404237d55c3df5c1180db71c23412fff629b2a06e5aaee4a2024dd2ff: Status 404 returned error can't find the container with id 26d4527404237d55c3df5c1180db71c23412fff629b2a06e5aaee4a2024dd2ff Dec 03 11:15:04 crc kubenswrapper[4646]: I1203 11:15:04.759289 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.003977 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d","Type":"ContainerStarted","Data":"c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2"} Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.004227 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="eaa2a4b9-4bb6-41d0-9f9e-601db339b10d" containerName="nova-metadata-metadata" containerID="cri-o://c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2" gracePeriod=30 Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.004225 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="eaa2a4b9-4bb6-41d0-9f9e-601db339b10d" containerName="nova-metadata-log" containerID="cri-o://6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed" gracePeriod=30 Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.007593 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"eb37f32e-cad3-4184-ad71-fd5cef58ca1f","Type":"ContainerStarted","Data":"26d4527404237d55c3df5c1180db71c23412fff629b2a06e5aaee4a2024dd2ff"} Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.009385 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6a5ea85e-a978-410c-90ed-2d1f95c2083c","Type":"ContainerStarted","Data":"f87302f841a41c588c01fa559ab8e6e1caa6b3850b2181edee2d9f7533a08692"} Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.015857 4646 generic.go:334] "Generic (PLEG): container finished" podID="52d1a48c-5c33-4ac1-95d6-709377ecbf30" containerID="2f928d4297817ebba8c477bdacbc5514c504bf4614e5c57b3764844ba04652ec" exitCode=0 Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.015951 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" event={"ID":"52d1a48c-5c33-4ac1-95d6-709377ecbf30","Type":"ContainerDied","Data":"2f928d4297817ebba8c477bdacbc5514c504bf4614e5c57b3764844ba04652ec"} Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.032270 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.416239506 podStartE2EDuration="8.032251543s" podCreationTimestamp="2025-12-03 11:14:57 +0000 UTC" firstStartedPulling="2025-12-03 11:14:58.762667172 +0000 UTC m=+1275.225723307" lastFinishedPulling="2025-12-03 11:15:03.378679209 +0000 UTC m=+1279.841735344" observedRunningTime="2025-12-03 11:15:05.029956318 +0000 UTC m=+1281.493012453" watchObservedRunningTime="2025-12-03 11:15:05.032251543 +0000 UTC m=+1281.495307678" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.069131 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.884635797 podStartE2EDuration="8.069110792s" podCreationTimestamp="2025-12-03 11:14:57 +0000 UTC" firstStartedPulling="2025-12-03 11:14:58.226395741 +0000 UTC m=+1274.689451876" lastFinishedPulling="2025-12-03 11:15:03.410870736 +0000 UTC m=+1279.873926871" observedRunningTime="2025-12-03 11:15:05.063305798 +0000 UTC m=+1281.526361933" watchObservedRunningTime="2025-12-03 11:15:05.069110792 +0000 UTC m=+1281.532166937" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.632255 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.670767 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.676564 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-config-data\") pod \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.676743 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-combined-ca-bundle\") pod \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.676825 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-sg-core-conf-yaml\") pod \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.676958 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-log-httpd\") pod \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.677052 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-run-httpd\") pod \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.677143 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-scripts\") pod \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.677256 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgrhd\" (UniqueName: \"kubernetes.io/projected/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-kube-api-access-hgrhd\") pod \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\" (UID: \"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c\") " Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.684059 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-kube-api-access-hgrhd" (OuterVolumeSpecName: "kube-api-access-hgrhd") pod "a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" (UID: "a7bddeee-2832-4e2d-9cd9-0cc05f781a2c"). InnerVolumeSpecName "kube-api-access-hgrhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.684675 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" (UID: "a7bddeee-2832-4e2d-9cd9-0cc05f781a2c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.685386 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" (UID: "a7bddeee-2832-4e2d-9cd9-0cc05f781a2c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.693224 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-scripts" (OuterVolumeSpecName: "scripts") pod "a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" (UID: "a7bddeee-2832-4e2d-9cd9-0cc05f781a2c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.744940 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" (UID: "a7bddeee-2832-4e2d-9cd9-0cc05f781a2c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.778505 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-logs\") pod \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.778684 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-config-data\") pod \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.778770 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp7qc\" (UniqueName: \"kubernetes.io/projected/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-kube-api-access-dp7qc\") pod \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.778812 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-combined-ca-bundle\") pod \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\" (UID: \"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d\") " Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.779144 4646 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.779154 4646 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.779162 4646 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.779172 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.779180 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgrhd\" (UniqueName: \"kubernetes.io/projected/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-kube-api-access-hgrhd\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.788255 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-logs" (OuterVolumeSpecName: "logs") pod "eaa2a4b9-4bb6-41d0-9f9e-601db339b10d" (UID: "eaa2a4b9-4bb6-41d0-9f9e-601db339b10d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.835775 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-kube-api-access-dp7qc" (OuterVolumeSpecName: "kube-api-access-dp7qc") pod "eaa2a4b9-4bb6-41d0-9f9e-601db339b10d" (UID: "eaa2a4b9-4bb6-41d0-9f9e-601db339b10d"). InnerVolumeSpecName "kube-api-access-dp7qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.842440 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eaa2a4b9-4bb6-41d0-9f9e-601db339b10d" (UID: "eaa2a4b9-4bb6-41d0-9f9e-601db339b10d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.854714 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" (UID: "a7bddeee-2832-4e2d-9cd9-0cc05f781a2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.865950 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-config-data" (OuterVolumeSpecName: "config-data") pod "eaa2a4b9-4bb6-41d0-9f9e-601db339b10d" (UID: "eaa2a4b9-4bb6-41d0-9f9e-601db339b10d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.874249 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86ef1c10-7d50-4089-a463-676380821fe3" path="/var/lib/kubelet/pods/86ef1c10-7d50-4089-a463-676380821fe3/volumes" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.886183 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp7qc\" (UniqueName: \"kubernetes.io/projected/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-kube-api-access-dp7qc\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.886208 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.886221 4646 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-logs\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.886231 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.886243 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.922518 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-config-data" (OuterVolumeSpecName: "config-data") pod "a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" (UID: "a7bddeee-2832-4e2d-9cd9-0cc05f781a2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:05 crc kubenswrapper[4646]: I1203 11:15:05.987447 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.027924 4646 generic.go:334] "Generic (PLEG): container finished" podID="eaa2a4b9-4bb6-41d0-9f9e-601db339b10d" containerID="c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2" exitCode=0 Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.027955 4646 generic.go:334] "Generic (PLEG): container finished" podID="eaa2a4b9-4bb6-41d0-9f9e-601db339b10d" containerID="6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed" exitCode=143 Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.027997 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d","Type":"ContainerDied","Data":"c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2"} Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.028029 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d","Type":"ContainerDied","Data":"6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed"} Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.028044 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eaa2a4b9-4bb6-41d0-9f9e-601db339b10d","Type":"ContainerDied","Data":"375cb185f3f881c959bef2486f122960b66bf539dd7325cc778d111e3faadade"} Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.028061 4646 scope.go:117] "RemoveContainer" containerID="c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.028196 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.039957 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"eb37f32e-cad3-4184-ad71-fd5cef58ca1f","Type":"ContainerStarted","Data":"cb7489187c1afb95c35d988641e2a83e21d9acd554cbc89b635c2e3fd84af649"} Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.041227 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.051899 4646 generic.go:334] "Generic (PLEG): container finished" podID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerID="de8d9e339b358ec44d4e0dad1b2be6569b0bd7ca298d68021356ca4e5add294f" exitCode=0 Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.052784 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.055461 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c","Type":"ContainerDied","Data":"de8d9e339b358ec44d4e0dad1b2be6569b0bd7ca298d68021356ca4e5add294f"} Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.055501 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7bddeee-2832-4e2d-9cd9-0cc05f781a2c","Type":"ContainerDied","Data":"ab006f93874712803df7ce50bcc8329e46d664df28287b0e9393ddf6024f81f1"} Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.068533 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.080558 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.105618 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:06 crc kubenswrapper[4646]: E1203 11:15:06.106282 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="ceilometer-central-agent" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.106418 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="ceilometer-central-agent" Dec 03 11:15:06 crc kubenswrapper[4646]: E1203 11:15:06.106485 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa2a4b9-4bb6-41d0-9f9e-601db339b10d" containerName="nova-metadata-metadata" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.106562 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa2a4b9-4bb6-41d0-9f9e-601db339b10d" containerName="nova-metadata-metadata" Dec 03 11:15:06 crc kubenswrapper[4646]: E1203 11:15:06.106658 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa2a4b9-4bb6-41d0-9f9e-601db339b10d" containerName="nova-metadata-log" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.106734 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa2a4b9-4bb6-41d0-9f9e-601db339b10d" containerName="nova-metadata-log" Dec 03 11:15:06 crc kubenswrapper[4646]: E1203 11:15:06.106818 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="proxy-httpd" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.106877 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="proxy-httpd" Dec 03 11:15:06 crc kubenswrapper[4646]: E1203 11:15:06.106948 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="ceilometer-notification-agent" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.107000 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="ceilometer-notification-agent" Dec 03 11:15:06 crc kubenswrapper[4646]: E1203 11:15:06.107069 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="sg-core" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.107133 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="sg-core" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.107388 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="sg-core" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.107455 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa2a4b9-4bb6-41d0-9f9e-601db339b10d" containerName="nova-metadata-metadata" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.107533 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="ceilometer-notification-agent" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.107810 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="ceilometer-central-agent" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.107867 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa2a4b9-4bb6-41d0-9f9e-601db339b10d" containerName="nova-metadata-log" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.107934 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" containerName="proxy-httpd" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.108979 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.123075 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.124170 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.132663 4646 scope.go:117] "RemoveContainer" containerID="6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.147064 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.189248 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.7898374989999999 podStartE2EDuration="2.189225613s" podCreationTimestamp="2025-12-03 11:15:04 +0000 UTC" firstStartedPulling="2025-12-03 11:15:04.766228147 +0000 UTC m=+1281.229284282" lastFinishedPulling="2025-12-03 11:15:05.165616261 +0000 UTC m=+1281.628672396" observedRunningTime="2025-12-03 11:15:06.108063186 +0000 UTC m=+1282.571119321" watchObservedRunningTime="2025-12-03 11:15:06.189225613 +0000 UTC m=+1282.652281748" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.193296 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.193437 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-config-data\") pod \"nova-metadata-0\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.193471 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.193508 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-logs\") pod \"nova-metadata-0\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.193525 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxsv7\" (UniqueName: \"kubernetes.io/projected/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-kube-api-access-xxsv7\") pod \"nova-metadata-0\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.232711 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.252413 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.275707 4646 scope.go:117] "RemoveContainer" containerID="c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2" Dec 03 11:15:06 crc kubenswrapper[4646]: E1203 11:15:06.277032 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2\": container with ID starting with c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2 not found: ID does not exist" containerID="c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.279076 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2"} err="failed to get container status \"c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2\": rpc error: code = NotFound desc = could not find container \"c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2\": container with ID starting with c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2 not found: ID does not exist" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.279102 4646 scope.go:117] "RemoveContainer" containerID="6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed" Dec 03 11:15:06 crc kubenswrapper[4646]: E1203 11:15:06.280734 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed\": container with ID starting with 6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed not found: ID does not exist" containerID="6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.280755 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed"} err="failed to get container status \"6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed\": rpc error: code = NotFound desc = could not find container \"6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed\": container with ID starting with 6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed not found: ID does not exist" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.280769 4646 scope.go:117] "RemoveContainer" containerID="c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.283848 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2"} err="failed to get container status \"c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2\": rpc error: code = NotFound desc = could not find container \"c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2\": container with ID starting with c930f422d9ff20c03b69a47ace6515fb816093935b1ad58c24aea815ca6bf2a2 not found: ID does not exist" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.283885 4646 scope.go:117] "RemoveContainer" containerID="6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.284163 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed"} err="failed to get container status \"6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed\": rpc error: code = NotFound desc = could not find container \"6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed\": container with ID starting with 6451cf6368be2047795520f17c3bc2fe576189ac54b88ff677c0c8b32c3007ed not found: ID does not exist" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.284179 4646 scope.go:117] "RemoveContainer" containerID="21b24b41604473c38dad81d0cffa3e5fa60589f2862a7460eb04b4992f372315" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.296261 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.296306 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-logs\") pod \"nova-metadata-0\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.296327 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxsv7\" (UniqueName: \"kubernetes.io/projected/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-kube-api-access-xxsv7\") pod \"nova-metadata-0\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.296422 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.296484 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-config-data\") pod \"nova-metadata-0\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.302725 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-logs\") pod \"nova-metadata-0\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.304957 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-config-data\") pod \"nova-metadata-0\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.311426 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.313493 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.313592 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.314555 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.321346 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.321521 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.323806 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.329644 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxsv7\" (UniqueName: \"kubernetes.io/projected/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-kube-api-access-xxsv7\") pod \"nova-metadata-0\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.333773 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.370476 4646 scope.go:117] "RemoveContainer" containerID="00190acf33ab3e604c60184a8c7350d5560756bab9bc79ad84634d8c99d767ff" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.405968 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lzdp\" (UniqueName: \"kubernetes.io/projected/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-kube-api-access-2lzdp\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.406027 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.406090 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-run-httpd\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.406110 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-config-data\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.406125 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.406154 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-scripts\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.406168 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.406196 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-log-httpd\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.422779 4646 scope.go:117] "RemoveContainer" containerID="de8d9e339b358ec44d4e0dad1b2be6569b0bd7ca298d68021356ca4e5add294f" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.445789 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.487500 4646 scope.go:117] "RemoveContainer" containerID="750d3b08a3b61d3cd698166b661c653dc1dde85b2a143951b403d255dfecbf8a" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.516381 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lzdp\" (UniqueName: \"kubernetes.io/projected/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-kube-api-access-2lzdp\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.516711 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.516777 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-run-httpd\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.516796 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-config-data\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.516814 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.516840 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-scripts\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.516854 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.516884 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-log-httpd\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.517238 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-log-httpd\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.519050 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-run-httpd\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.525428 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-scripts\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.526301 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.528026 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.528386 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.529079 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-config-data\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.548048 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lzdp\" (UniqueName: \"kubernetes.io/projected/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-kube-api-access-2lzdp\") pod \"ceilometer-0\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.592013 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.600134 4646 scope.go:117] "RemoveContainer" containerID="21b24b41604473c38dad81d0cffa3e5fa60589f2862a7460eb04b4992f372315" Dec 03 11:15:06 crc kubenswrapper[4646]: E1203 11:15:06.603323 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21b24b41604473c38dad81d0cffa3e5fa60589f2862a7460eb04b4992f372315\": container with ID starting with 21b24b41604473c38dad81d0cffa3e5fa60589f2862a7460eb04b4992f372315 not found: ID does not exist" containerID="21b24b41604473c38dad81d0cffa3e5fa60589f2862a7460eb04b4992f372315" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.603378 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21b24b41604473c38dad81d0cffa3e5fa60589f2862a7460eb04b4992f372315"} err="failed to get container status \"21b24b41604473c38dad81d0cffa3e5fa60589f2862a7460eb04b4992f372315\": rpc error: code = NotFound desc = could not find container \"21b24b41604473c38dad81d0cffa3e5fa60589f2862a7460eb04b4992f372315\": container with ID starting with 21b24b41604473c38dad81d0cffa3e5fa60589f2862a7460eb04b4992f372315 not found: ID does not exist" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.603401 4646 scope.go:117] "RemoveContainer" containerID="00190acf33ab3e604c60184a8c7350d5560756bab9bc79ad84634d8c99d767ff" Dec 03 11:15:06 crc kubenswrapper[4646]: E1203 11:15:06.604436 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00190acf33ab3e604c60184a8c7350d5560756bab9bc79ad84634d8c99d767ff\": container with ID starting with 00190acf33ab3e604c60184a8c7350d5560756bab9bc79ad84634d8c99d767ff not found: ID does not exist" containerID="00190acf33ab3e604c60184a8c7350d5560756bab9bc79ad84634d8c99d767ff" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.604457 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00190acf33ab3e604c60184a8c7350d5560756bab9bc79ad84634d8c99d767ff"} err="failed to get container status \"00190acf33ab3e604c60184a8c7350d5560756bab9bc79ad84634d8c99d767ff\": rpc error: code = NotFound desc = could not find container \"00190acf33ab3e604c60184a8c7350d5560756bab9bc79ad84634d8c99d767ff\": container with ID starting with 00190acf33ab3e604c60184a8c7350d5560756bab9bc79ad84634d8c99d767ff not found: ID does not exist" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.604511 4646 scope.go:117] "RemoveContainer" containerID="de8d9e339b358ec44d4e0dad1b2be6569b0bd7ca298d68021356ca4e5add294f" Dec 03 11:15:06 crc kubenswrapper[4646]: E1203 11:15:06.606126 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de8d9e339b358ec44d4e0dad1b2be6569b0bd7ca298d68021356ca4e5add294f\": container with ID starting with de8d9e339b358ec44d4e0dad1b2be6569b0bd7ca298d68021356ca4e5add294f not found: ID does not exist" containerID="de8d9e339b358ec44d4e0dad1b2be6569b0bd7ca298d68021356ca4e5add294f" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.606182 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de8d9e339b358ec44d4e0dad1b2be6569b0bd7ca298d68021356ca4e5add294f"} err="failed to get container status \"de8d9e339b358ec44d4e0dad1b2be6569b0bd7ca298d68021356ca4e5add294f\": rpc error: code = NotFound desc = could not find container \"de8d9e339b358ec44d4e0dad1b2be6569b0bd7ca298d68021356ca4e5add294f\": container with ID starting with de8d9e339b358ec44d4e0dad1b2be6569b0bd7ca298d68021356ca4e5add294f not found: ID does not exist" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.606203 4646 scope.go:117] "RemoveContainer" containerID="750d3b08a3b61d3cd698166b661c653dc1dde85b2a143951b403d255dfecbf8a" Dec 03 11:15:06 crc kubenswrapper[4646]: E1203 11:15:06.610435 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"750d3b08a3b61d3cd698166b661c653dc1dde85b2a143951b403d255dfecbf8a\": container with ID starting with 750d3b08a3b61d3cd698166b661c653dc1dde85b2a143951b403d255dfecbf8a not found: ID does not exist" containerID="750d3b08a3b61d3cd698166b661c653dc1dde85b2a143951b403d255dfecbf8a" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.610478 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"750d3b08a3b61d3cd698166b661c653dc1dde85b2a143951b403d255dfecbf8a"} err="failed to get container status \"750d3b08a3b61d3cd698166b661c653dc1dde85b2a143951b403d255dfecbf8a\": rpc error: code = NotFound desc = could not find container \"750d3b08a3b61d3cd698166b661c653dc1dde85b2a143951b403d255dfecbf8a\": container with ID starting with 750d3b08a3b61d3cd698166b661c653dc1dde85b2a143951b403d255dfecbf8a not found: ID does not exist" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.721631 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52d1a48c-5c33-4ac1-95d6-709377ecbf30-config-volume\") pod \"52d1a48c-5c33-4ac1-95d6-709377ecbf30\" (UID: \"52d1a48c-5c33-4ac1-95d6-709377ecbf30\") " Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.721680 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52d1a48c-5c33-4ac1-95d6-709377ecbf30-secret-volume\") pod \"52d1a48c-5c33-4ac1-95d6-709377ecbf30\" (UID: \"52d1a48c-5c33-4ac1-95d6-709377ecbf30\") " Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.721770 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4hhs\" (UniqueName: \"kubernetes.io/projected/52d1a48c-5c33-4ac1-95d6-709377ecbf30-kube-api-access-s4hhs\") pod \"52d1a48c-5c33-4ac1-95d6-709377ecbf30\" (UID: \"52d1a48c-5c33-4ac1-95d6-709377ecbf30\") " Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.725570 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52d1a48c-5c33-4ac1-95d6-709377ecbf30-config-volume" (OuterVolumeSpecName: "config-volume") pod "52d1a48c-5c33-4ac1-95d6-709377ecbf30" (UID: "52d1a48c-5c33-4ac1-95d6-709377ecbf30"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.726109 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.728640 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52d1a48c-5c33-4ac1-95d6-709377ecbf30-kube-api-access-s4hhs" (OuterVolumeSpecName: "kube-api-access-s4hhs") pod "52d1a48c-5c33-4ac1-95d6-709377ecbf30" (UID: "52d1a48c-5c33-4ac1-95d6-709377ecbf30"). InnerVolumeSpecName "kube-api-access-s4hhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.730600 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52d1a48c-5c33-4ac1-95d6-709377ecbf30-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "52d1a48c-5c33-4ac1-95d6-709377ecbf30" (UID: "52d1a48c-5c33-4ac1-95d6-709377ecbf30"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.825513 4646 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52d1a48c-5c33-4ac1-95d6-709377ecbf30-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.825547 4646 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52d1a48c-5c33-4ac1-95d6-709377ecbf30-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.825560 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4hhs\" (UniqueName: \"kubernetes.io/projected/52d1a48c-5c33-4ac1-95d6-709377ecbf30-kube-api-access-s4hhs\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:06 crc kubenswrapper[4646]: I1203 11:15:06.981120 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:06 crc kubenswrapper[4646]: W1203 11:15:06.984783 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4a93ebd_b2ed_4eb1_b4f0_056551484d97.slice/crio-1cfc23ed48899267f3c3476e914b0bc036ed62985af4b2939cff2cf43817b1fc WatchSource:0}: Error finding container 1cfc23ed48899267f3c3476e914b0bc036ed62985af4b2939cff2cf43817b1fc: Status 404 returned error can't find the container with id 1cfc23ed48899267f3c3476e914b0bc036ed62985af4b2939cff2cf43817b1fc Dec 03 11:15:07 crc kubenswrapper[4646]: I1203 11:15:07.075680 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" event={"ID":"52d1a48c-5c33-4ac1-95d6-709377ecbf30","Type":"ContainerDied","Data":"41f61111f5d3e15a2a3110ed73910349c93276e942081a6bd2f471355e1ddf56"} Dec 03 11:15:07 crc kubenswrapper[4646]: I1203 11:15:07.075975 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41f61111f5d3e15a2a3110ed73910349c93276e942081a6bd2f471355e1ddf56" Dec 03 11:15:07 crc kubenswrapper[4646]: I1203 11:15:07.076033 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412675-66h54" Dec 03 11:15:07 crc kubenswrapper[4646]: I1203 11:15:07.079995 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f4a93ebd-b2ed-4eb1-b4f0-056551484d97","Type":"ContainerStarted","Data":"1cfc23ed48899267f3c3476e914b0bc036ed62985af4b2939cff2cf43817b1fc"} Dec 03 11:15:07 crc kubenswrapper[4646]: I1203 11:15:07.239424 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:15:07 crc kubenswrapper[4646]: I1203 11:15:07.279449 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="86ef1c10-7d50-4089-a463-676380821fe3" containerName="kube-state-metrics" probeResult="failure" output="Get \"http://10.217.0.103:8081/readyz\": dial tcp 10.217.0.103:8081: i/o timeout (Client.Timeout exceeded while awaiting headers)" Dec 03 11:15:07 crc kubenswrapper[4646]: I1203 11:15:07.449647 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 11:15:07 crc kubenswrapper[4646]: I1203 11:15:07.449688 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 11:15:07 crc kubenswrapper[4646]: I1203 11:15:07.483559 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 11:15:07 crc kubenswrapper[4646]: I1203 11:15:07.483616 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 11:15:07 crc kubenswrapper[4646]: I1203 11:15:07.505223 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 11:15:07 crc kubenswrapper[4646]: I1203 11:15:07.857509 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7bddeee-2832-4e2d-9cd9-0cc05f781a2c" path="/var/lib/kubelet/pods/a7bddeee-2832-4e2d-9cd9-0cc05f781a2c/volumes" Dec 03 11:15:07 crc kubenswrapper[4646]: I1203 11:15:07.858217 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaa2a4b9-4bb6-41d0-9f9e-601db339b10d" path="/var/lib/kubelet/pods/eaa2a4b9-4bb6-41d0-9f9e-601db339b10d/volumes" Dec 03 11:15:07 crc kubenswrapper[4646]: I1203 11:15:07.869436 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:08 crc kubenswrapper[4646]: I1203 11:15:08.091696 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8","Type":"ContainerStarted","Data":"226337a866312357ba40f1a0db7ec0bf96a65f4eb46ed3994e41dd985a1ba1d2"} Dec 03 11:15:08 crc kubenswrapper[4646]: I1203 11:15:08.091742 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8","Type":"ContainerStarted","Data":"e67a8b26b0b0935ee3595921bb7aae50b42a1b9a59acf01321507fd827649d2b"} Dec 03 11:15:08 crc kubenswrapper[4646]: I1203 11:15:08.094485 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f4a93ebd-b2ed-4eb1-b4f0-056551484d97","Type":"ContainerStarted","Data":"c7f73a8f8a25022df700acfac552ee2e1cfdeaaabc1460e4238fdba6879d43b9"} Dec 03 11:15:08 crc kubenswrapper[4646]: I1203 11:15:08.094518 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f4a93ebd-b2ed-4eb1-b4f0-056551484d97","Type":"ContainerStarted","Data":"fbab94241e17153dddb0c4064eb472396c4ea4018b66270aba71d370e7387203"} Dec 03 11:15:08 crc kubenswrapper[4646]: I1203 11:15:08.115216 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.115196602 podStartE2EDuration="2.115196602s" podCreationTimestamp="2025-12-03 11:15:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:15:08.110892651 +0000 UTC m=+1284.573948786" watchObservedRunningTime="2025-12-03 11:15:08.115196602 +0000 UTC m=+1284.578252737" Dec 03 11:15:08 crc kubenswrapper[4646]: I1203 11:15:08.145726 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 11:15:08 crc kubenswrapper[4646]: I1203 11:15:08.566606 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6a5ea85e-a978-410c-90ed-2d1f95c2083c" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.168:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 11:15:08 crc kubenswrapper[4646]: I1203 11:15:08.566848 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6a5ea85e-a978-410c-90ed-2d1f95c2083c" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.168:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 11:15:09 crc kubenswrapper[4646]: I1203 11:15:09.119704 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8","Type":"ContainerStarted","Data":"fe7acf9d5da8040edf3fcb08b5e2fbd194d97f1eda5ac4d210ccb34497e92fd9"} Dec 03 11:15:10 crc kubenswrapper[4646]: I1203 11:15:10.128100 4646 generic.go:334] "Generic (PLEG): container finished" podID="f37ef5fc-6976-46cf-87ee-6c3d8acf75a8" containerID="59ab2024684ff6d3c8ff7ab17b5b5e02ba1aef3d3411b4759799dca1d9e408b4" exitCode=0 Dec 03 11:15:10 crc kubenswrapper[4646]: I1203 11:15:10.128543 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vx27k" event={"ID":"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8","Type":"ContainerDied","Data":"59ab2024684ff6d3c8ff7ab17b5b5e02ba1aef3d3411b4759799dca1d9e408b4"} Dec 03 11:15:10 crc kubenswrapper[4646]: I1203 11:15:10.131594 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8","Type":"ContainerStarted","Data":"c0554e3a5c4c523da89bee4d069ea3535af17ebaf6da6238b7f045c1eec1ddf3"} Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.146094 4646 generic.go:334] "Generic (PLEG): container finished" podID="56efccd7-660f-4052-afa3-dab6a63e93a2" containerID="55f494c40970f92c45b900e8e523f7be3167fa07ff2296242dd792044a06cf32" exitCode=0 Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.146158 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-j4gb5" event={"ID":"56efccd7-660f-4052-afa3-dab6a63e93a2","Type":"ContainerDied","Data":"55f494c40970f92c45b900e8e523f7be3167fa07ff2296242dd792044a06cf32"} Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.150840 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8","Type":"ContainerStarted","Data":"03e018aa4169175ce602de9e4767fff2300ce0f1dbc626845c717f2c9bde9c0c"} Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.151160 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.211790 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.58774393 podStartE2EDuration="5.211769306s" podCreationTimestamp="2025-12-03 11:15:06 +0000 UTC" firstStartedPulling="2025-12-03 11:15:07.244794846 +0000 UTC m=+1283.707850981" lastFinishedPulling="2025-12-03 11:15:10.868820222 +0000 UTC m=+1287.331876357" observedRunningTime="2025-12-03 11:15:11.204360247 +0000 UTC m=+1287.667416392" watchObservedRunningTime="2025-12-03 11:15:11.211769306 +0000 UTC m=+1287.674825451" Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.447238 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.447612 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.580386 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.727705 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-scripts\") pod \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.727797 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrlnz\" (UniqueName: \"kubernetes.io/projected/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-kube-api-access-vrlnz\") pod \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.727965 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-config-data\") pod \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.727999 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-combined-ca-bundle\") pod \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.733673 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-scripts" (OuterVolumeSpecName: "scripts") pod "f37ef5fc-6976-46cf-87ee-6c3d8acf75a8" (UID: "f37ef5fc-6976-46cf-87ee-6c3d8acf75a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.733978 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-kube-api-access-vrlnz" (OuterVolumeSpecName: "kube-api-access-vrlnz") pod "f37ef5fc-6976-46cf-87ee-6c3d8acf75a8" (UID: "f37ef5fc-6976-46cf-87ee-6c3d8acf75a8"). InnerVolumeSpecName "kube-api-access-vrlnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:11 crc kubenswrapper[4646]: E1203 11:15:11.753073 4646 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-config-data podName:f37ef5fc-6976-46cf-87ee-6c3d8acf75a8 nodeName:}" failed. No retries permitted until 2025-12-03 11:15:12.253046277 +0000 UTC m=+1288.716102412 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-config-data") pod "f37ef5fc-6976-46cf-87ee-6c3d8acf75a8" (UID: "f37ef5fc-6976-46cf-87ee-6c3d8acf75a8") : error deleting /var/lib/kubelet/pods/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8/volume-subpaths: remove /var/lib/kubelet/pods/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8/volume-subpaths: no such file or directory Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.758436 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f37ef5fc-6976-46cf-87ee-6c3d8acf75a8" (UID: "f37ef5fc-6976-46cf-87ee-6c3d8acf75a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.830386 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.830419 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:11 crc kubenswrapper[4646]: I1203 11:15:11.830429 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrlnz\" (UniqueName: \"kubernetes.io/projected/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-kube-api-access-vrlnz\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.160190 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vx27k" Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.160311 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vx27k" event={"ID":"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8","Type":"ContainerDied","Data":"0f65f2400fcd031a401f817e1d15bc59e590cff42aeefd6bc2ef5e3e3144c406"} Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.160348 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f65f2400fcd031a401f817e1d15bc59e590cff42aeefd6bc2ef5e3e3144c406" Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.340955 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-config-data\") pod \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\" (UID: \"f37ef5fc-6976-46cf-87ee-6c3d8acf75a8\") " Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.351502 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-config-data" (OuterVolumeSpecName: "config-data") pod "f37ef5fc-6976-46cf-87ee-6c3d8acf75a8" (UID: "f37ef5fc-6976-46cf-87ee-6c3d8acf75a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.373667 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.373917 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6a5ea85e-a978-410c-90ed-2d1f95c2083c" containerName="nova-api-log" containerID="cri-o://e235ce8e8d49e6703d27bca078986c6c7514030f02ade612e8e7a1d92de748ee" gracePeriod=30 Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.374233 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6a5ea85e-a978-410c-90ed-2d1f95c2083c" containerName="nova-api-api" containerID="cri-o://f87302f841a41c588c01fa559ab8e6e1caa6b3850b2181edee2d9f7533a08692" gracePeriod=30 Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.441953 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.442165 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="46aeaecf-b036-48cb-8b33-04c134aa5b8b" containerName="nova-scheduler-scheduler" containerID="cri-o://37ffdf894aa4e90c1ca1185efe46685416dfa0b2dab568f761aca9d242c6e5f2" gracePeriod=30 Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.447795 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:12 crc kubenswrapper[4646]: E1203 11:15:12.459455 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="37ffdf894aa4e90c1ca1185efe46685416dfa0b2dab568f761aca9d242c6e5f2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 11:15:12 crc kubenswrapper[4646]: E1203 11:15:12.467476 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="37ffdf894aa4e90c1ca1185efe46685416dfa0b2dab568f761aca9d242c6e5f2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 11:15:12 crc kubenswrapper[4646]: E1203 11:15:12.474463 4646 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="37ffdf894aa4e90c1ca1185efe46685416dfa0b2dab568f761aca9d242c6e5f2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 11:15:12 crc kubenswrapper[4646]: E1203 11:15:12.474542 4646 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="46aeaecf-b036-48cb-8b33-04c134aa5b8b" containerName="nova-scheduler-scheduler" Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.484300 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.484536 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f4a93ebd-b2ed-4eb1-b4f0-056551484d97" containerName="nova-metadata-log" containerID="cri-o://fbab94241e17153dddb0c4064eb472396c4ea4018b66270aba71d370e7387203" gracePeriod=30 Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.484923 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f4a93ebd-b2ed-4eb1-b4f0-056551484d97" containerName="nova-metadata-metadata" containerID="cri-o://c7f73a8f8a25022df700acfac552ee2e1cfdeaaabc1460e4238fdba6879d43b9" gracePeriod=30 Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.690727 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.772414 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-combined-ca-bundle\") pod \"56efccd7-660f-4052-afa3-dab6a63e93a2\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.776127 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-scripts\") pod \"56efccd7-660f-4052-afa3-dab6a63e93a2\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.776292 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twl52\" (UniqueName: \"kubernetes.io/projected/56efccd7-660f-4052-afa3-dab6a63e93a2-kube-api-access-twl52\") pod \"56efccd7-660f-4052-afa3-dab6a63e93a2\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.776354 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-config-data\") pod \"56efccd7-660f-4052-afa3-dab6a63e93a2\" (UID: \"56efccd7-660f-4052-afa3-dab6a63e93a2\") " Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.782399 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56efccd7-660f-4052-afa3-dab6a63e93a2-kube-api-access-twl52" (OuterVolumeSpecName: "kube-api-access-twl52") pod "56efccd7-660f-4052-afa3-dab6a63e93a2" (UID: "56efccd7-660f-4052-afa3-dab6a63e93a2"). InnerVolumeSpecName "kube-api-access-twl52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.783128 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-scripts" (OuterVolumeSpecName: "scripts") pod "56efccd7-660f-4052-afa3-dab6a63e93a2" (UID: "56efccd7-660f-4052-afa3-dab6a63e93a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.801152 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56efccd7-660f-4052-afa3-dab6a63e93a2" (UID: "56efccd7-660f-4052-afa3-dab6a63e93a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.805518 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-config-data" (OuterVolumeSpecName: "config-data") pod "56efccd7-660f-4052-afa3-dab6a63e93a2" (UID: "56efccd7-660f-4052-afa3-dab6a63e93a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.878480 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twl52\" (UniqueName: \"kubernetes.io/projected/56efccd7-660f-4052-afa3-dab6a63e93a2-kube-api-access-twl52\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.878515 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.878524 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.878536 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56efccd7-660f-4052-afa3-dab6a63e93a2-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:12 crc kubenswrapper[4646]: I1203 11:15:12.975525 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.041360 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-ddhcl"] Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.041911 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" podUID="196f11c5-52de-42c3-ac45-dbaad2588b1d" containerName="dnsmasq-dns" containerID="cri-o://d1bde7db1f0b6ab39d82c1db108e699eb7edf462ede886515bb85cd726915308" gracePeriod=10 Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.172796 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-j4gb5" event={"ID":"56efccd7-660f-4052-afa3-dab6a63e93a2","Type":"ContainerDied","Data":"7c17282edadde051def412cb1c96c070106ee7946d7cfe385174590435a13797"} Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.172832 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c17282edadde051def412cb1c96c070106ee7946d7cfe385174590435a13797" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.172885 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-j4gb5" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.186346 4646 generic.go:334] "Generic (PLEG): container finished" podID="6a5ea85e-a978-410c-90ed-2d1f95c2083c" containerID="e235ce8e8d49e6703d27bca078986c6c7514030f02ade612e8e7a1d92de748ee" exitCode=143 Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.186437 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6a5ea85e-a978-410c-90ed-2d1f95c2083c","Type":"ContainerDied","Data":"e235ce8e8d49e6703d27bca078986c6c7514030f02ade612e8e7a1d92de748ee"} Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.225244 4646 generic.go:334] "Generic (PLEG): container finished" podID="46aeaecf-b036-48cb-8b33-04c134aa5b8b" containerID="37ffdf894aa4e90c1ca1185efe46685416dfa0b2dab568f761aca9d242c6e5f2" exitCode=0 Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.225362 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"46aeaecf-b036-48cb-8b33-04c134aa5b8b","Type":"ContainerDied","Data":"37ffdf894aa4e90c1ca1185efe46685416dfa0b2dab568f761aca9d242c6e5f2"} Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.254323 4646 generic.go:334] "Generic (PLEG): container finished" podID="f4a93ebd-b2ed-4eb1-b4f0-056551484d97" containerID="c7f73a8f8a25022df700acfac552ee2e1cfdeaaabc1460e4238fdba6879d43b9" exitCode=0 Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.254365 4646 generic.go:334] "Generic (PLEG): container finished" podID="f4a93ebd-b2ed-4eb1-b4f0-056551484d97" containerID="fbab94241e17153dddb0c4064eb472396c4ea4018b66270aba71d370e7387203" exitCode=143 Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.254386 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f4a93ebd-b2ed-4eb1-b4f0-056551484d97","Type":"ContainerDied","Data":"c7f73a8f8a25022df700acfac552ee2e1cfdeaaabc1460e4238fdba6879d43b9"} Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.254412 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f4a93ebd-b2ed-4eb1-b4f0-056551484d97","Type":"ContainerDied","Data":"fbab94241e17153dddb0c4064eb472396c4ea4018b66270aba71d370e7387203"} Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.313565 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 11:15:13 crc kubenswrapper[4646]: E1203 11:15:13.314026 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56efccd7-660f-4052-afa3-dab6a63e93a2" containerName="nova-cell1-conductor-db-sync" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.314045 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="56efccd7-660f-4052-afa3-dab6a63e93a2" containerName="nova-cell1-conductor-db-sync" Dec 03 11:15:13 crc kubenswrapper[4646]: E1203 11:15:13.314088 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52d1a48c-5c33-4ac1-95d6-709377ecbf30" containerName="collect-profiles" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.314096 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="52d1a48c-5c33-4ac1-95d6-709377ecbf30" containerName="collect-profiles" Dec 03 11:15:13 crc kubenswrapper[4646]: E1203 11:15:13.314118 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f37ef5fc-6976-46cf-87ee-6c3d8acf75a8" containerName="nova-manage" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.314126 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f37ef5fc-6976-46cf-87ee-6c3d8acf75a8" containerName="nova-manage" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.314352 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="56efccd7-660f-4052-afa3-dab6a63e93a2" containerName="nova-cell1-conductor-db-sync" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.314374 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="f37ef5fc-6976-46cf-87ee-6c3d8acf75a8" containerName="nova-manage" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.314390 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="52d1a48c-5c33-4ac1-95d6-709377ecbf30" containerName="collect-profiles" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.315134 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.325242 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.339931 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.413647 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82149567-e81e-4056-829c-a854bee435c5-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"82149567-e81e-4056-829c-a854bee435c5\") " pod="openstack/nova-cell1-conductor-0" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.413701 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzlwr\" (UniqueName: \"kubernetes.io/projected/82149567-e81e-4056-829c-a854bee435c5-kube-api-access-lzlwr\") pod \"nova-cell1-conductor-0\" (UID: \"82149567-e81e-4056-829c-a854bee435c5\") " pod="openstack/nova-cell1-conductor-0" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.413758 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82149567-e81e-4056-829c-a854bee435c5-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"82149567-e81e-4056-829c-a854bee435c5\") " pod="openstack/nova-cell1-conductor-0" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.517008 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzlwr\" (UniqueName: \"kubernetes.io/projected/82149567-e81e-4056-829c-a854bee435c5-kube-api-access-lzlwr\") pod \"nova-cell1-conductor-0\" (UID: \"82149567-e81e-4056-829c-a854bee435c5\") " pod="openstack/nova-cell1-conductor-0" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.517075 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82149567-e81e-4056-829c-a854bee435c5-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"82149567-e81e-4056-829c-a854bee435c5\") " pod="openstack/nova-cell1-conductor-0" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.517205 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82149567-e81e-4056-829c-a854bee435c5-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"82149567-e81e-4056-829c-a854bee435c5\") " pod="openstack/nova-cell1-conductor-0" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.537990 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82149567-e81e-4056-829c-a854bee435c5-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"82149567-e81e-4056-829c-a854bee435c5\") " pod="openstack/nova-cell1-conductor-0" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.541151 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82149567-e81e-4056-829c-a854bee435c5-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"82149567-e81e-4056-829c-a854bee435c5\") " pod="openstack/nova-cell1-conductor-0" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.545992 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzlwr\" (UniqueName: \"kubernetes.io/projected/82149567-e81e-4056-829c-a854bee435c5-kube-api-access-lzlwr\") pod \"nova-cell1-conductor-0\" (UID: \"82149567-e81e-4056-829c-a854bee435c5\") " pod="openstack/nova-cell1-conductor-0" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.600301 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.633430 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.644199 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.721189 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6dtp\" (UniqueName: \"kubernetes.io/projected/46aeaecf-b036-48cb-8b33-04c134aa5b8b-kube-api-access-c6dtp\") pod \"46aeaecf-b036-48cb-8b33-04c134aa5b8b\" (UID: \"46aeaecf-b036-48cb-8b33-04c134aa5b8b\") " Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.721587 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46aeaecf-b036-48cb-8b33-04c134aa5b8b-config-data\") pod \"46aeaecf-b036-48cb-8b33-04c134aa5b8b\" (UID: \"46aeaecf-b036-48cb-8b33-04c134aa5b8b\") " Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.721773 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46aeaecf-b036-48cb-8b33-04c134aa5b8b-combined-ca-bundle\") pod \"46aeaecf-b036-48cb-8b33-04c134aa5b8b\" (UID: \"46aeaecf-b036-48cb-8b33-04c134aa5b8b\") " Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.761535 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46aeaecf-b036-48cb-8b33-04c134aa5b8b-kube-api-access-c6dtp" (OuterVolumeSpecName: "kube-api-access-c6dtp") pod "46aeaecf-b036-48cb-8b33-04c134aa5b8b" (UID: "46aeaecf-b036-48cb-8b33-04c134aa5b8b"). InnerVolumeSpecName "kube-api-access-c6dtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.811387 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46aeaecf-b036-48cb-8b33-04c134aa5b8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46aeaecf-b036-48cb-8b33-04c134aa5b8b" (UID: "46aeaecf-b036-48cb-8b33-04c134aa5b8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.815507 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46aeaecf-b036-48cb-8b33-04c134aa5b8b-config-data" (OuterVolumeSpecName: "config-data") pod "46aeaecf-b036-48cb-8b33-04c134aa5b8b" (UID: "46aeaecf-b036-48cb-8b33-04c134aa5b8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.823250 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-config-data\") pod \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.823297 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-nova-metadata-tls-certs\") pod \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.823348 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxsv7\" (UniqueName: \"kubernetes.io/projected/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-kube-api-access-xxsv7\") pod \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.823384 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-logs\") pod \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.823407 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-combined-ca-bundle\") pod \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\" (UID: \"f4a93ebd-b2ed-4eb1-b4f0-056551484d97\") " Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.823823 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6dtp\" (UniqueName: \"kubernetes.io/projected/46aeaecf-b036-48cb-8b33-04c134aa5b8b-kube-api-access-c6dtp\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.823836 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46aeaecf-b036-48cb-8b33-04c134aa5b8b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.823847 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46aeaecf-b036-48cb-8b33-04c134aa5b8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.827113 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-logs" (OuterVolumeSpecName: "logs") pod "f4a93ebd-b2ed-4eb1-b4f0-056551484d97" (UID: "f4a93ebd-b2ed-4eb1-b4f0-056551484d97"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.833142 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-kube-api-access-xxsv7" (OuterVolumeSpecName: "kube-api-access-xxsv7") pod "f4a93ebd-b2ed-4eb1-b4f0-056551484d97" (UID: "f4a93ebd-b2ed-4eb1-b4f0-056551484d97"). InnerVolumeSpecName "kube-api-access-xxsv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.844954 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.869380 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-config-data" (OuterVolumeSpecName: "config-data") pod "f4a93ebd-b2ed-4eb1-b4f0-056551484d97" (UID: "f4a93ebd-b2ed-4eb1-b4f0-056551484d97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.922172 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4a93ebd-b2ed-4eb1-b4f0-056551484d97" (UID: "f4a93ebd-b2ed-4eb1-b4f0-056551484d97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.925227 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.925243 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxsv7\" (UniqueName: \"kubernetes.io/projected/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-kube-api-access-xxsv7\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.925253 4646 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-logs\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.925263 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:13 crc kubenswrapper[4646]: I1203 11:15:13.941549 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f4a93ebd-b2ed-4eb1-b4f0-056551484d97" (UID: "f4a93ebd-b2ed-4eb1-b4f0-056551484d97"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.028588 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-config\") pod \"196f11c5-52de-42c3-ac45-dbaad2588b1d\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.028679 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kpvs\" (UniqueName: \"kubernetes.io/projected/196f11c5-52de-42c3-ac45-dbaad2588b1d-kube-api-access-5kpvs\") pod \"196f11c5-52de-42c3-ac45-dbaad2588b1d\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.028735 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-dns-svc\") pod \"196f11c5-52de-42c3-ac45-dbaad2588b1d\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.028820 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-ovsdbserver-nb\") pod \"196f11c5-52de-42c3-ac45-dbaad2588b1d\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.028960 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-ovsdbserver-sb\") pod \"196f11c5-52de-42c3-ac45-dbaad2588b1d\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.035347 4646 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4a93ebd-b2ed-4eb1-b4f0-056551484d97-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.050171 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/196f11c5-52de-42c3-ac45-dbaad2588b1d-kube-api-access-5kpvs" (OuterVolumeSpecName: "kube-api-access-5kpvs") pod "196f11c5-52de-42c3-ac45-dbaad2588b1d" (UID: "196f11c5-52de-42c3-ac45-dbaad2588b1d"). InnerVolumeSpecName "kube-api-access-5kpvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.110224 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "196f11c5-52de-42c3-ac45-dbaad2588b1d" (UID: "196f11c5-52de-42c3-ac45-dbaad2588b1d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.136639 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "196f11c5-52de-42c3-ac45-dbaad2588b1d" (UID: "196f11c5-52de-42c3-ac45-dbaad2588b1d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.136826 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-ovsdbserver-nb\") pod \"196f11c5-52de-42c3-ac45-dbaad2588b1d\" (UID: \"196f11c5-52de-42c3-ac45-dbaad2588b1d\") " Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.137260 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kpvs\" (UniqueName: \"kubernetes.io/projected/196f11c5-52de-42c3-ac45-dbaad2588b1d-kube-api-access-5kpvs\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.137278 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:14 crc kubenswrapper[4646]: W1203 11:15:14.137443 4646 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/196f11c5-52de-42c3-ac45-dbaad2588b1d/volumes/kubernetes.io~configmap/ovsdbserver-nb Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.137458 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "196f11c5-52de-42c3-ac45-dbaad2588b1d" (UID: "196f11c5-52de-42c3-ac45-dbaad2588b1d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.197162 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-config" (OuterVolumeSpecName: "config") pod "196f11c5-52de-42c3-ac45-dbaad2588b1d" (UID: "196f11c5-52de-42c3-ac45-dbaad2588b1d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.238839 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.238883 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.242271 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "196f11c5-52de-42c3-ac45-dbaad2588b1d" (UID: "196f11c5-52de-42c3-ac45-dbaad2588b1d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.274576 4646 generic.go:334] "Generic (PLEG): container finished" podID="196f11c5-52de-42c3-ac45-dbaad2588b1d" containerID="d1bde7db1f0b6ab39d82c1db108e699eb7edf462ede886515bb85cd726915308" exitCode=0 Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.274648 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" event={"ID":"196f11c5-52de-42c3-ac45-dbaad2588b1d","Type":"ContainerDied","Data":"d1bde7db1f0b6ab39d82c1db108e699eb7edf462ede886515bb85cd726915308"} Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.274680 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" event={"ID":"196f11c5-52de-42c3-ac45-dbaad2588b1d","Type":"ContainerDied","Data":"058092684cf790a447fd7368d9528a6d442431a79d2415a55c97334a3371408d"} Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.274698 4646 scope.go:117] "RemoveContainer" containerID="d1bde7db1f0b6ab39d82c1db108e699eb7edf462ede886515bb85cd726915308" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.274839 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-ddhcl" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.284874 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"46aeaecf-b036-48cb-8b33-04c134aa5b8b","Type":"ContainerDied","Data":"d7c232213893431e48fc76e9e4afbdac1bd54f910a204fd232cf0e0cacda89c5"} Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.284979 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.293093 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f4a93ebd-b2ed-4eb1-b4f0-056551484d97","Type":"ContainerDied","Data":"1cfc23ed48899267f3c3476e914b0bc036ed62985af4b2939cff2cf43817b1fc"} Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.293143 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.330520 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-ddhcl"] Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.330649 4646 scope.go:117] "RemoveContainer" containerID="516220b559295f4349fd056e4928f5d6622bdb69d1c8724ef0b206c458265b36" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.340970 4646 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/196f11c5-52de-42c3-ac45-dbaad2588b1d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.364606 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-ddhcl"] Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.380042 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.393768 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.405855 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.408489 4646 scope.go:117] "RemoveContainer" containerID="d1bde7db1f0b6ab39d82c1db108e699eb7edf462ede886515bb85cd726915308" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.414286 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 11:15:14 crc kubenswrapper[4646]: E1203 11:15:14.415447 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1bde7db1f0b6ab39d82c1db108e699eb7edf462ede886515bb85cd726915308\": container with ID starting with d1bde7db1f0b6ab39d82c1db108e699eb7edf462ede886515bb85cd726915308 not found: ID does not exist" containerID="d1bde7db1f0b6ab39d82c1db108e699eb7edf462ede886515bb85cd726915308" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.415488 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1bde7db1f0b6ab39d82c1db108e699eb7edf462ede886515bb85cd726915308"} err="failed to get container status \"d1bde7db1f0b6ab39d82c1db108e699eb7edf462ede886515bb85cd726915308\": rpc error: code = NotFound desc = could not find container \"d1bde7db1f0b6ab39d82c1db108e699eb7edf462ede886515bb85cd726915308\": container with ID starting with d1bde7db1f0b6ab39d82c1db108e699eb7edf462ede886515bb85cd726915308 not found: ID does not exist" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.415514 4646 scope.go:117] "RemoveContainer" containerID="516220b559295f4349fd056e4928f5d6622bdb69d1c8724ef0b206c458265b36" Dec 03 11:15:14 crc kubenswrapper[4646]: E1203 11:15:14.415741 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a93ebd-b2ed-4eb1-b4f0-056551484d97" containerName="nova-metadata-log" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.415768 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a93ebd-b2ed-4eb1-b4f0-056551484d97" containerName="nova-metadata-log" Dec 03 11:15:14 crc kubenswrapper[4646]: E1203 11:15:14.415800 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="196f11c5-52de-42c3-ac45-dbaad2588b1d" containerName="dnsmasq-dns" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.415809 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="196f11c5-52de-42c3-ac45-dbaad2588b1d" containerName="dnsmasq-dns" Dec 03 11:15:14 crc kubenswrapper[4646]: E1203 11:15:14.415939 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"516220b559295f4349fd056e4928f5d6622bdb69d1c8724ef0b206c458265b36\": container with ID starting with 516220b559295f4349fd056e4928f5d6622bdb69d1c8724ef0b206c458265b36 not found: ID does not exist" containerID="516220b559295f4349fd056e4928f5d6622bdb69d1c8724ef0b206c458265b36" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.415977 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"516220b559295f4349fd056e4928f5d6622bdb69d1c8724ef0b206c458265b36"} err="failed to get container status \"516220b559295f4349fd056e4928f5d6622bdb69d1c8724ef0b206c458265b36\": rpc error: code = NotFound desc = could not find container \"516220b559295f4349fd056e4928f5d6622bdb69d1c8724ef0b206c458265b36\": container with ID starting with 516220b559295f4349fd056e4928f5d6622bdb69d1c8724ef0b206c458265b36 not found: ID does not exist" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.416002 4646 scope.go:117] "RemoveContainer" containerID="37ffdf894aa4e90c1ca1185efe46685416dfa0b2dab568f761aca9d242c6e5f2" Dec 03 11:15:14 crc kubenswrapper[4646]: E1203 11:15:14.415822 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46aeaecf-b036-48cb-8b33-04c134aa5b8b" containerName="nova-scheduler-scheduler" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.416224 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="46aeaecf-b036-48cb-8b33-04c134aa5b8b" containerName="nova-scheduler-scheduler" Dec 03 11:15:14 crc kubenswrapper[4646]: E1203 11:15:14.416242 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="196f11c5-52de-42c3-ac45-dbaad2588b1d" containerName="init" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.416250 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="196f11c5-52de-42c3-ac45-dbaad2588b1d" containerName="init" Dec 03 11:15:14 crc kubenswrapper[4646]: E1203 11:15:14.416272 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a93ebd-b2ed-4eb1-b4f0-056551484d97" containerName="nova-metadata-metadata" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.416280 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a93ebd-b2ed-4eb1-b4f0-056551484d97" containerName="nova-metadata-metadata" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.416745 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="46aeaecf-b036-48cb-8b33-04c134aa5b8b" containerName="nova-scheduler-scheduler" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.416768 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4a93ebd-b2ed-4eb1-b4f0-056551484d97" containerName="nova-metadata-metadata" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.416779 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="196f11c5-52de-42c3-ac45-dbaad2588b1d" containerName="dnsmasq-dns" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.416792 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4a93ebd-b2ed-4eb1-b4f0-056551484d97" containerName="nova-metadata-log" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.418706 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.426413 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.439578 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.461481 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.461754 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.471872 4646 scope.go:117] "RemoveContainer" containerID="c7f73a8f8a25022df700acfac552ee2e1cfdeaaabc1460e4238fdba6879d43b9" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.482095 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.483690 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.486681 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.486839 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.507930 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.526553 4646 scope.go:117] "RemoveContainer" containerID="fbab94241e17153dddb0c4064eb472396c4ea4018b66270aba71d370e7387203" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.545182 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnf75\" (UniqueName: \"kubernetes.io/projected/a7a75658-d6ae-4875-a787-0b3e49697499-kube-api-access-nnf75\") pod \"nova-scheduler-0\" (UID: \"a7a75658-d6ae-4875-a787-0b3e49697499\") " pod="openstack/nova-scheduler-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.545248 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7a75658-d6ae-4875-a787-0b3e49697499-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a7a75658-d6ae-4875-a787-0b3e49697499\") " pod="openstack/nova-scheduler-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.545298 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7a75658-d6ae-4875-a787-0b3e49697499-config-data\") pod \"nova-scheduler-0\" (UID: \"a7a75658-d6ae-4875-a787-0b3e49697499\") " pod="openstack/nova-scheduler-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.562206 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.646791 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnf75\" (UniqueName: \"kubernetes.io/projected/a7a75658-d6ae-4875-a787-0b3e49697499-kube-api-access-nnf75\") pod \"nova-scheduler-0\" (UID: \"a7a75658-d6ae-4875-a787-0b3e49697499\") " pod="openstack/nova-scheduler-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.646846 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.646922 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7a75658-d6ae-4875-a787-0b3e49697499-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a7a75658-d6ae-4875-a787-0b3e49697499\") " pod="openstack/nova-scheduler-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.647010 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7a75658-d6ae-4875-a787-0b3e49697499-config-data\") pod \"nova-scheduler-0\" (UID: \"a7a75658-d6ae-4875-a787-0b3e49697499\") " pod="openstack/nova-scheduler-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.647042 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.647079 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-config-data\") pod \"nova-metadata-0\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.647097 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfmnr\" (UniqueName: \"kubernetes.io/projected/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-kube-api-access-qfmnr\") pod \"nova-metadata-0\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.647113 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-logs\") pod \"nova-metadata-0\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.679771 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7a75658-d6ae-4875-a787-0b3e49697499-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a7a75658-d6ae-4875-a787-0b3e49697499\") " pod="openstack/nova-scheduler-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.685323 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7a75658-d6ae-4875-a787-0b3e49697499-config-data\") pod \"nova-scheduler-0\" (UID: \"a7a75658-d6ae-4875-a787-0b3e49697499\") " pod="openstack/nova-scheduler-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.692821 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnf75\" (UniqueName: \"kubernetes.io/projected/a7a75658-d6ae-4875-a787-0b3e49697499-kube-api-access-nnf75\") pod \"nova-scheduler-0\" (UID: \"a7a75658-d6ae-4875-a787-0b3e49697499\") " pod="openstack/nova-scheduler-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.748045 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.748137 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.748155 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-config-data\") pod \"nova-metadata-0\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.748174 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfmnr\" (UniqueName: \"kubernetes.io/projected/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-kube-api-access-qfmnr\") pod \"nova-metadata-0\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.748191 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-logs\") pod \"nova-metadata-0\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.748549 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-logs\") pod \"nova-metadata-0\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.750481 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.760855 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-config-data\") pod \"nova-metadata-0\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.771028 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.771054 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.784358 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfmnr\" (UniqueName: \"kubernetes.io/projected/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-kube-api-access-qfmnr\") pod \"nova-metadata-0\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " pod="openstack/nova-metadata-0" Dec 03 11:15:14 crc kubenswrapper[4646]: I1203 11:15:14.816983 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 11:15:15 crc kubenswrapper[4646]: I1203 11:15:15.305778 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"82149567-e81e-4056-829c-a854bee435c5","Type":"ContainerStarted","Data":"a082ad29f528b5626e848f7ada31681877437df52adff9a0156b54e7ee2590de"} Dec 03 11:15:15 crc kubenswrapper[4646]: I1203 11:15:15.306046 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"82149567-e81e-4056-829c-a854bee435c5","Type":"ContainerStarted","Data":"c53df246638530cf93957a64672f32ab0741b235df017cfec89387ac2191996a"} Dec 03 11:15:15 crc kubenswrapper[4646]: I1203 11:15:15.306157 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 11:15:15 crc kubenswrapper[4646]: I1203 11:15:15.326130 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.326108247 podStartE2EDuration="2.326108247s" podCreationTimestamp="2025-12-03 11:15:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:15:15.322086154 +0000 UTC m=+1291.785142299" watchObservedRunningTime="2025-12-03 11:15:15.326108247 +0000 UTC m=+1291.789164392" Dec 03 11:15:15 crc kubenswrapper[4646]: W1203 11:15:15.343035 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5dc2d79_d598_4b4d_80fd_4ac11612b32f.slice/crio-644b3b6d9b1afb184b487eda04cb7bda8784fe805939261c15c79e4d3c254981 WatchSource:0}: Error finding container 644b3b6d9b1afb184b487eda04cb7bda8784fe805939261c15c79e4d3c254981: Status 404 returned error can't find the container with id 644b3b6d9b1afb184b487eda04cb7bda8784fe805939261c15c79e4d3c254981 Dec 03 11:15:15 crc kubenswrapper[4646]: I1203 11:15:15.359081 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:15 crc kubenswrapper[4646]: I1203 11:15:15.369706 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 11:15:15 crc kubenswrapper[4646]: I1203 11:15:15.864271 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="196f11c5-52de-42c3-ac45-dbaad2588b1d" path="/var/lib/kubelet/pods/196f11c5-52de-42c3-ac45-dbaad2588b1d/volumes" Dec 03 11:15:15 crc kubenswrapper[4646]: I1203 11:15:15.866445 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46aeaecf-b036-48cb-8b33-04c134aa5b8b" path="/var/lib/kubelet/pods/46aeaecf-b036-48cb-8b33-04c134aa5b8b/volumes" Dec 03 11:15:15 crc kubenswrapper[4646]: I1203 11:15:15.867001 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4a93ebd-b2ed-4eb1-b4f0-056551484d97" path="/var/lib/kubelet/pods/f4a93ebd-b2ed-4eb1-b4f0-056551484d97/volumes" Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.325443 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5dc2d79-d598-4b4d-80fd-4ac11612b32f","Type":"ContainerStarted","Data":"1b476e700d6473b5090594015176b8ef1b20eab1116d439a30d346cb91ea20d6"} Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.325491 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5dc2d79-d598-4b4d-80fd-4ac11612b32f","Type":"ContainerStarted","Data":"5e09820f691f7079067c490be8272a91f0bb4eec739e409bb9604ab4d889a4b4"} Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.325507 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5dc2d79-d598-4b4d-80fd-4ac11612b32f","Type":"ContainerStarted","Data":"644b3b6d9b1afb184b487eda04cb7bda8784fe805939261c15c79e4d3c254981"} Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.328216 4646 generic.go:334] "Generic (PLEG): container finished" podID="6a5ea85e-a978-410c-90ed-2d1f95c2083c" containerID="f87302f841a41c588c01fa559ab8e6e1caa6b3850b2181edee2d9f7533a08692" exitCode=0 Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.328260 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6a5ea85e-a978-410c-90ed-2d1f95c2083c","Type":"ContainerDied","Data":"f87302f841a41c588c01fa559ab8e6e1caa6b3850b2181edee2d9f7533a08692"} Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.329512 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a7a75658-d6ae-4875-a787-0b3e49697499","Type":"ContainerStarted","Data":"5265a31fb663da564a289470d89f019e099cf570cad86d4d793f454e0ede77fa"} Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.329541 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a7a75658-d6ae-4875-a787-0b3e49697499","Type":"ContainerStarted","Data":"57d8de499733cc2e0dc409603316ec1598b6fc351ac8a1e5cc2429f4b5c5a4ff"} Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.383202 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.3831799719999998 podStartE2EDuration="2.383179972s" podCreationTimestamp="2025-12-03 11:15:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:15:16.35723485 +0000 UTC m=+1292.820290985" watchObservedRunningTime="2025-12-03 11:15:16.383179972 +0000 UTC m=+1292.846236117" Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.401614 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.40159695 podStartE2EDuration="2.40159695s" podCreationTimestamp="2025-12-03 11:15:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:15:16.379038685 +0000 UTC m=+1292.842094830" watchObservedRunningTime="2025-12-03 11:15:16.40159695 +0000 UTC m=+1292.864653105" Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.809497 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.888528 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckhjj\" (UniqueName: \"kubernetes.io/projected/6a5ea85e-a978-410c-90ed-2d1f95c2083c-kube-api-access-ckhjj\") pod \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.888573 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5ea85e-a978-410c-90ed-2d1f95c2083c-combined-ca-bundle\") pod \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.888679 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5ea85e-a978-410c-90ed-2d1f95c2083c-logs\") pod \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.888725 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5ea85e-a978-410c-90ed-2d1f95c2083c-config-data\") pod \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\" (UID: \"6a5ea85e-a978-410c-90ed-2d1f95c2083c\") " Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.890045 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a5ea85e-a978-410c-90ed-2d1f95c2083c-logs" (OuterVolumeSpecName: "logs") pod "6a5ea85e-a978-410c-90ed-2d1f95c2083c" (UID: "6a5ea85e-a978-410c-90ed-2d1f95c2083c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.909398 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a5ea85e-a978-410c-90ed-2d1f95c2083c-kube-api-access-ckhjj" (OuterVolumeSpecName: "kube-api-access-ckhjj") pod "6a5ea85e-a978-410c-90ed-2d1f95c2083c" (UID: "6a5ea85e-a978-410c-90ed-2d1f95c2083c"). InnerVolumeSpecName "kube-api-access-ckhjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.918868 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5ea85e-a978-410c-90ed-2d1f95c2083c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a5ea85e-a978-410c-90ed-2d1f95c2083c" (UID: "6a5ea85e-a978-410c-90ed-2d1f95c2083c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.934798 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5ea85e-a978-410c-90ed-2d1f95c2083c-config-data" (OuterVolumeSpecName: "config-data") pod "6a5ea85e-a978-410c-90ed-2d1f95c2083c" (UID: "6a5ea85e-a978-410c-90ed-2d1f95c2083c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.991018 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckhjj\" (UniqueName: \"kubernetes.io/projected/6a5ea85e-a978-410c-90ed-2d1f95c2083c-kube-api-access-ckhjj\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.991043 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5ea85e-a978-410c-90ed-2d1f95c2083c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.991052 4646 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5ea85e-a978-410c-90ed-2d1f95c2083c-logs\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:16 crc kubenswrapper[4646]: I1203 11:15:16.991063 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5ea85e-a978-410c-90ed-2d1f95c2083c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.342679 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.344991 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6a5ea85e-a978-410c-90ed-2d1f95c2083c","Type":"ContainerDied","Data":"f3909bbd4d68645b10bccf95a049c2ef45966b55f54913ca0c44985b50f43c2b"} Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.345085 4646 scope.go:117] "RemoveContainer" containerID="f87302f841a41c588c01fa559ab8e6e1caa6b3850b2181edee2d9f7533a08692" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.369324 4646 scope.go:117] "RemoveContainer" containerID="e235ce8e8d49e6703d27bca078986c6c7514030f02ade612e8e7a1d92de748ee" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.407480 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.436609 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.454800 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:17 crc kubenswrapper[4646]: E1203 11:15:17.455263 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5ea85e-a978-410c-90ed-2d1f95c2083c" containerName="nova-api-log" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.455285 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5ea85e-a978-410c-90ed-2d1f95c2083c" containerName="nova-api-log" Dec 03 11:15:17 crc kubenswrapper[4646]: E1203 11:15:17.455311 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5ea85e-a978-410c-90ed-2d1f95c2083c" containerName="nova-api-api" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.455320 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5ea85e-a978-410c-90ed-2d1f95c2083c" containerName="nova-api-api" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.455581 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5ea85e-a978-410c-90ed-2d1f95c2083c" containerName="nova-api-api" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.455604 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5ea85e-a978-410c-90ed-2d1f95c2083c" containerName="nova-api-log" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.457517 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.473998 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.475221 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.600650 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prbcm\" (UniqueName: \"kubernetes.io/projected/82837c9d-805c-43f1-90cf-16f4bd77c469-kube-api-access-prbcm\") pod \"nova-api-0\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " pod="openstack/nova-api-0" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.600782 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82837c9d-805c-43f1-90cf-16f4bd77c469-logs\") pod \"nova-api-0\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " pod="openstack/nova-api-0" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.600848 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82837c9d-805c-43f1-90cf-16f4bd77c469-config-data\") pod \"nova-api-0\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " pod="openstack/nova-api-0" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.600974 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82837c9d-805c-43f1-90cf-16f4bd77c469-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " pod="openstack/nova-api-0" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.702647 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82837c9d-805c-43f1-90cf-16f4bd77c469-config-data\") pod \"nova-api-0\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " pod="openstack/nova-api-0" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.702742 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82837c9d-805c-43f1-90cf-16f4bd77c469-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " pod="openstack/nova-api-0" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.703454 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prbcm\" (UniqueName: \"kubernetes.io/projected/82837c9d-805c-43f1-90cf-16f4bd77c469-kube-api-access-prbcm\") pod \"nova-api-0\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " pod="openstack/nova-api-0" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.703604 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82837c9d-805c-43f1-90cf-16f4bd77c469-logs\") pod \"nova-api-0\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " pod="openstack/nova-api-0" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.703978 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82837c9d-805c-43f1-90cf-16f4bd77c469-logs\") pod \"nova-api-0\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " pod="openstack/nova-api-0" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.706746 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82837c9d-805c-43f1-90cf-16f4bd77c469-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " pod="openstack/nova-api-0" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.708173 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82837c9d-805c-43f1-90cf-16f4bd77c469-config-data\") pod \"nova-api-0\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " pod="openstack/nova-api-0" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.728899 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prbcm\" (UniqueName: \"kubernetes.io/projected/82837c9d-805c-43f1-90cf-16f4bd77c469-kube-api-access-prbcm\") pod \"nova-api-0\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " pod="openstack/nova-api-0" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.787851 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 11:15:17 crc kubenswrapper[4646]: I1203 11:15:17.863088 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a5ea85e-a978-410c-90ed-2d1f95c2083c" path="/var/lib/kubelet/pods/6a5ea85e-a978-410c-90ed-2d1f95c2083c/volumes" Dec 03 11:15:18 crc kubenswrapper[4646]: I1203 11:15:18.280752 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:18 crc kubenswrapper[4646]: I1203 11:15:18.349964 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"82837c9d-805c-43f1-90cf-16f4bd77c469","Type":"ContainerStarted","Data":"12efbc22dbacf165d04839bb1ce1503b00185dd11f965876ff7bba2526edab20"} Dec 03 11:15:19 crc kubenswrapper[4646]: I1203 11:15:19.380719 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"82837c9d-805c-43f1-90cf-16f4bd77c469","Type":"ContainerStarted","Data":"c44b0a9d8d679da495cfad1c2f86c0ec38e16d3ad75edd95b20c8a3b7bc5513c"} Dec 03 11:15:19 crc kubenswrapper[4646]: I1203 11:15:19.381050 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"82837c9d-805c-43f1-90cf-16f4bd77c469","Type":"ContainerStarted","Data":"bd02bbc3ac3a06c5406acb8db6c4c07a7df6b6a2a98709fa794b2398e5e9dbd1"} Dec 03 11:15:19 crc kubenswrapper[4646]: I1203 11:15:19.410304 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.410286658 podStartE2EDuration="2.410286658s" podCreationTimestamp="2025-12-03 11:15:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:15:19.404367311 +0000 UTC m=+1295.867423446" watchObservedRunningTime="2025-12-03 11:15:19.410286658 +0000 UTC m=+1295.873342793" Dec 03 11:15:19 crc kubenswrapper[4646]: I1203 11:15:19.752027 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 11:15:19 crc kubenswrapper[4646]: I1203 11:15:19.817784 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 11:15:19 crc kubenswrapper[4646]: I1203 11:15:19.817844 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 11:15:23 crc kubenswrapper[4646]: I1203 11:15:23.703033 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 11:15:24 crc kubenswrapper[4646]: I1203 11:15:24.751227 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 11:15:24 crc kubenswrapper[4646]: I1203 11:15:24.794803 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 11:15:24 crc kubenswrapper[4646]: I1203 11:15:24.820699 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 11:15:24 crc kubenswrapper[4646]: I1203 11:15:24.820754 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 11:15:25 crc kubenswrapper[4646]: I1203 11:15:25.459296 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 11:15:25 crc kubenswrapper[4646]: I1203 11:15:25.840608 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e5dc2d79-d598-4b4d-80fd-4ac11612b32f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.179:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 11:15:25 crc kubenswrapper[4646]: I1203 11:15:25.840634 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e5dc2d79-d598-4b4d-80fd-4ac11612b32f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.179:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 11:15:27 crc kubenswrapper[4646]: I1203 11:15:27.789204 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 11:15:27 crc kubenswrapper[4646]: I1203 11:15:27.789268 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 11:15:28 crc kubenswrapper[4646]: I1203 11:15:28.872554 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="82837c9d-805c-43f1-90cf-16f4bd77c469" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.180:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 11:15:28 crc kubenswrapper[4646]: I1203 11:15:28.872554 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="82837c9d-805c-43f1-90cf-16f4bd77c469" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.180:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.416897 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.513404 4646 generic.go:334] "Generic (PLEG): container finished" podID="97a4238d-da32-427d-91cc-16d2df65221f" containerID="8397ac3883bdb4300287cf24c3dc1bee449b192f8df04dfee3d59fabfb9ad03c" exitCode=137 Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.513454 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"97a4238d-da32-427d-91cc-16d2df65221f","Type":"ContainerDied","Data":"8397ac3883bdb4300287cf24c3dc1bee449b192f8df04dfee3d59fabfb9ad03c"} Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.513725 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"97a4238d-da32-427d-91cc-16d2df65221f","Type":"ContainerDied","Data":"5b94909c2e3206d64d2ce2df3f8bc35db5abbf83975362b4a8c0fe96e9362f01"} Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.513751 4646 scope.go:117] "RemoveContainer" containerID="8397ac3883bdb4300287cf24c3dc1bee449b192f8df04dfee3d59fabfb9ad03c" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.513469 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.515605 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwvk8\" (UniqueName: \"kubernetes.io/projected/97a4238d-da32-427d-91cc-16d2df65221f-kube-api-access-wwvk8\") pod \"97a4238d-da32-427d-91cc-16d2df65221f\" (UID: \"97a4238d-da32-427d-91cc-16d2df65221f\") " Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.515653 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a4238d-da32-427d-91cc-16d2df65221f-config-data\") pod \"97a4238d-da32-427d-91cc-16d2df65221f\" (UID: \"97a4238d-da32-427d-91cc-16d2df65221f\") " Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.516864 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a4238d-da32-427d-91cc-16d2df65221f-combined-ca-bundle\") pod \"97a4238d-da32-427d-91cc-16d2df65221f\" (UID: \"97a4238d-da32-427d-91cc-16d2df65221f\") " Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.522643 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97a4238d-da32-427d-91cc-16d2df65221f-kube-api-access-wwvk8" (OuterVolumeSpecName: "kube-api-access-wwvk8") pod "97a4238d-da32-427d-91cc-16d2df65221f" (UID: "97a4238d-da32-427d-91cc-16d2df65221f"). InnerVolumeSpecName "kube-api-access-wwvk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.543456 4646 scope.go:117] "RemoveContainer" containerID="8397ac3883bdb4300287cf24c3dc1bee449b192f8df04dfee3d59fabfb9ad03c" Dec 03 11:15:34 crc kubenswrapper[4646]: E1203 11:15:34.544297 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8397ac3883bdb4300287cf24c3dc1bee449b192f8df04dfee3d59fabfb9ad03c\": container with ID starting with 8397ac3883bdb4300287cf24c3dc1bee449b192f8df04dfee3d59fabfb9ad03c not found: ID does not exist" containerID="8397ac3883bdb4300287cf24c3dc1bee449b192f8df04dfee3d59fabfb9ad03c" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.544360 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8397ac3883bdb4300287cf24c3dc1bee449b192f8df04dfee3d59fabfb9ad03c"} err="failed to get container status \"8397ac3883bdb4300287cf24c3dc1bee449b192f8df04dfee3d59fabfb9ad03c\": rpc error: code = NotFound desc = could not find container \"8397ac3883bdb4300287cf24c3dc1bee449b192f8df04dfee3d59fabfb9ad03c\": container with ID starting with 8397ac3883bdb4300287cf24c3dc1bee449b192f8df04dfee3d59fabfb9ad03c not found: ID does not exist" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.554995 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97a4238d-da32-427d-91cc-16d2df65221f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "97a4238d-da32-427d-91cc-16d2df65221f" (UID: "97a4238d-da32-427d-91cc-16d2df65221f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.557784 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97a4238d-da32-427d-91cc-16d2df65221f-config-data" (OuterVolumeSpecName: "config-data") pod "97a4238d-da32-427d-91cc-16d2df65221f" (UID: "97a4238d-da32-427d-91cc-16d2df65221f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.619780 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwvk8\" (UniqueName: \"kubernetes.io/projected/97a4238d-da32-427d-91cc-16d2df65221f-kube-api-access-wwvk8\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.619822 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a4238d-da32-427d-91cc-16d2df65221f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.619833 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a4238d-da32-427d-91cc-16d2df65221f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.823396 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.829325 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.833386 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.896113 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.906573 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.935768 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 11:15:34 crc kubenswrapper[4646]: E1203 11:15:34.936183 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97a4238d-da32-427d-91cc-16d2df65221f" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.936206 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a4238d-da32-427d-91cc-16d2df65221f" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.936515 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="97a4238d-da32-427d-91cc-16d2df65221f" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.937189 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.940659 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.941407 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.941553 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 03 11:15:34 crc kubenswrapper[4646]: I1203 11:15:34.950072 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.027197 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a064e62-8fce-4cb7-9d19-e1e0542570c3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a064e62-8fce-4cb7-9d19-e1e0542570c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.027385 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a064e62-8fce-4cb7-9d19-e1e0542570c3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a064e62-8fce-4cb7-9d19-e1e0542570c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.027413 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fcdg\" (UniqueName: \"kubernetes.io/projected/9a064e62-8fce-4cb7-9d19-e1e0542570c3-kube-api-access-2fcdg\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a064e62-8fce-4cb7-9d19-e1e0542570c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.027543 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a064e62-8fce-4cb7-9d19-e1e0542570c3-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a064e62-8fce-4cb7-9d19-e1e0542570c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.027585 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a064e62-8fce-4cb7-9d19-e1e0542570c3-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a064e62-8fce-4cb7-9d19-e1e0542570c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.129328 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a064e62-8fce-4cb7-9d19-e1e0542570c3-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a064e62-8fce-4cb7-9d19-e1e0542570c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.129443 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a064e62-8fce-4cb7-9d19-e1e0542570c3-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a064e62-8fce-4cb7-9d19-e1e0542570c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.129494 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a064e62-8fce-4cb7-9d19-e1e0542570c3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a064e62-8fce-4cb7-9d19-e1e0542570c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.129562 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a064e62-8fce-4cb7-9d19-e1e0542570c3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a064e62-8fce-4cb7-9d19-e1e0542570c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.129601 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fcdg\" (UniqueName: \"kubernetes.io/projected/9a064e62-8fce-4cb7-9d19-e1e0542570c3-kube-api-access-2fcdg\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a064e62-8fce-4cb7-9d19-e1e0542570c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.138818 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a064e62-8fce-4cb7-9d19-e1e0542570c3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a064e62-8fce-4cb7-9d19-e1e0542570c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.143821 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a064e62-8fce-4cb7-9d19-e1e0542570c3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a064e62-8fce-4cb7-9d19-e1e0542570c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.146842 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a064e62-8fce-4cb7-9d19-e1e0542570c3-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a064e62-8fce-4cb7-9d19-e1e0542570c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.147763 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a064e62-8fce-4cb7-9d19-e1e0542570c3-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a064e62-8fce-4cb7-9d19-e1e0542570c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.163128 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fcdg\" (UniqueName: \"kubernetes.io/projected/9a064e62-8fce-4cb7-9d19-e1e0542570c3-kube-api-access-2fcdg\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a064e62-8fce-4cb7-9d19-e1e0542570c3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.258958 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.528883 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.757057 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 11:15:35 crc kubenswrapper[4646]: W1203 11:15:35.757826 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a064e62_8fce_4cb7_9d19_e1e0542570c3.slice/crio-cc43354b24769ccf6af1bed252d80dee41f28df68742c99f7925899c2ae0d3bf WatchSource:0}: Error finding container cc43354b24769ccf6af1bed252d80dee41f28df68742c99f7925899c2ae0d3bf: Status 404 returned error can't find the container with id cc43354b24769ccf6af1bed252d80dee41f28df68742c99f7925899c2ae0d3bf Dec 03 11:15:35 crc kubenswrapper[4646]: I1203 11:15:35.867982 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97a4238d-da32-427d-91cc-16d2df65221f" path="/var/lib/kubelet/pods/97a4238d-da32-427d-91cc-16d2df65221f/volumes" Dec 03 11:15:36 crc kubenswrapper[4646]: I1203 11:15:36.548064 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9a064e62-8fce-4cb7-9d19-e1e0542570c3","Type":"ContainerStarted","Data":"50c1037f611137cb1271b99fa1c91c211ad0f0e61256a94fee80d0f490d0cace"} Dec 03 11:15:36 crc kubenswrapper[4646]: I1203 11:15:36.548174 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9a064e62-8fce-4cb7-9d19-e1e0542570c3","Type":"ContainerStarted","Data":"cc43354b24769ccf6af1bed252d80dee41f28df68742c99f7925899c2ae0d3bf"} Dec 03 11:15:36 crc kubenswrapper[4646]: I1203 11:15:36.567308 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.56729145 podStartE2EDuration="2.56729145s" podCreationTimestamp="2025-12-03 11:15:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:15:36.564901193 +0000 UTC m=+1313.027957348" watchObservedRunningTime="2025-12-03 11:15:36.56729145 +0000 UTC m=+1313.030347585" Dec 03 11:15:36 crc kubenswrapper[4646]: I1203 11:15:36.736858 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 11:15:37 crc kubenswrapper[4646]: I1203 11:15:37.792550 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 11:15:37 crc kubenswrapper[4646]: I1203 11:15:37.793550 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 11:15:37 crc kubenswrapper[4646]: I1203 11:15:37.793823 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 11:15:37 crc kubenswrapper[4646]: I1203 11:15:37.793895 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 11:15:37 crc kubenswrapper[4646]: I1203 11:15:37.795980 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 11:15:37 crc kubenswrapper[4646]: I1203 11:15:37.800285 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.058995 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-5cknk"] Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.060647 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.077511 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92xcj\" (UniqueName: \"kubernetes.io/projected/14a94107-452f-4dc7-a96f-4cb1e0d455d4-kube-api-access-92xcj\") pod \"dnsmasq-dns-68d4b6d797-5cknk\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.077792 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-dns-svc\") pod \"dnsmasq-dns-68d4b6d797-5cknk\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.077988 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-config\") pod \"dnsmasq-dns-68d4b6d797-5cknk\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.078126 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-ovsdbserver-nb\") pod \"dnsmasq-dns-68d4b6d797-5cknk\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.078268 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-ovsdbserver-sb\") pod \"dnsmasq-dns-68d4b6d797-5cknk\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.102071 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-5cknk"] Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.182503 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-config\") pod \"dnsmasq-dns-68d4b6d797-5cknk\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.182558 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-ovsdbserver-nb\") pod \"dnsmasq-dns-68d4b6d797-5cknk\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.182606 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-ovsdbserver-sb\") pod \"dnsmasq-dns-68d4b6d797-5cknk\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.182681 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92xcj\" (UniqueName: \"kubernetes.io/projected/14a94107-452f-4dc7-a96f-4cb1e0d455d4-kube-api-access-92xcj\") pod \"dnsmasq-dns-68d4b6d797-5cknk\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.182707 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-dns-svc\") pod \"dnsmasq-dns-68d4b6d797-5cknk\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.183676 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-dns-svc\") pod \"dnsmasq-dns-68d4b6d797-5cknk\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.184228 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-config\") pod \"dnsmasq-dns-68d4b6d797-5cknk\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.184782 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-ovsdbserver-nb\") pod \"dnsmasq-dns-68d4b6d797-5cknk\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.185587 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-ovsdbserver-sb\") pod \"dnsmasq-dns-68d4b6d797-5cknk\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.209265 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92xcj\" (UniqueName: \"kubernetes.io/projected/14a94107-452f-4dc7-a96f-4cb1e0d455d4-kube-api-access-92xcj\") pod \"dnsmasq-dns-68d4b6d797-5cknk\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.384766 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:38 crc kubenswrapper[4646]: I1203 11:15:38.890713 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-5cknk"] Dec 03 11:15:39 crc kubenswrapper[4646]: I1203 11:15:39.595019 4646 generic.go:334] "Generic (PLEG): container finished" podID="14a94107-452f-4dc7-a96f-4cb1e0d455d4" containerID="215bb2e16524a7f844284d45b3e832b96911f027ca6b06a2230d5be443e886e8" exitCode=0 Dec 03 11:15:39 crc kubenswrapper[4646]: I1203 11:15:39.595075 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" event={"ID":"14a94107-452f-4dc7-a96f-4cb1e0d455d4","Type":"ContainerDied","Data":"215bb2e16524a7f844284d45b3e832b96911f027ca6b06a2230d5be443e886e8"} Dec 03 11:15:39 crc kubenswrapper[4646]: I1203 11:15:39.595398 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" event={"ID":"14a94107-452f-4dc7-a96f-4cb1e0d455d4","Type":"ContainerStarted","Data":"1130e383f5ebc2ab19de2432161f19994aeee5ed6043eb7fdacf3457d6c8dc00"} Dec 03 11:15:40 crc kubenswrapper[4646]: I1203 11:15:40.259208 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:40 crc kubenswrapper[4646]: I1203 11:15:40.565723 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:40 crc kubenswrapper[4646]: I1203 11:15:40.605046 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" event={"ID":"14a94107-452f-4dc7-a96f-4cb1e0d455d4","Type":"ContainerStarted","Data":"b3cd3b39c3c6c10f4ad806e0c9e17b7abbbd2895c57390aa3abfa02e54e78108"} Dec 03 11:15:40 crc kubenswrapper[4646]: I1203 11:15:40.605156 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="82837c9d-805c-43f1-90cf-16f4bd77c469" containerName="nova-api-log" containerID="cri-o://bd02bbc3ac3a06c5406acb8db6c4c07a7df6b6a2a98709fa794b2398e5e9dbd1" gracePeriod=30 Dec 03 11:15:40 crc kubenswrapper[4646]: I1203 11:15:40.605191 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="82837c9d-805c-43f1-90cf-16f4bd77c469" containerName="nova-api-api" containerID="cri-o://c44b0a9d8d679da495cfad1c2f86c0ec38e16d3ad75edd95b20c8a3b7bc5513c" gracePeriod=30 Dec 03 11:15:40 crc kubenswrapper[4646]: I1203 11:15:40.636843 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" podStartSLOduration=2.636818238 podStartE2EDuration="2.636818238s" podCreationTimestamp="2025-12-03 11:15:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:15:40.63224665 +0000 UTC m=+1317.095302795" watchObservedRunningTime="2025-12-03 11:15:40.636818238 +0000 UTC m=+1317.099874373" Dec 03 11:15:40 crc kubenswrapper[4646]: I1203 11:15:40.801940 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:15:40 crc kubenswrapper[4646]: I1203 11:15:40.802195 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="ceilometer-central-agent" containerID="cri-o://226337a866312357ba40f1a0db7ec0bf96a65f4eb46ed3994e41dd985a1ba1d2" gracePeriod=30 Dec 03 11:15:40 crc kubenswrapper[4646]: I1203 11:15:40.802232 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="sg-core" containerID="cri-o://c0554e3a5c4c523da89bee4d069ea3535af17ebaf6da6238b7f045c1eec1ddf3" gracePeriod=30 Dec 03 11:15:40 crc kubenswrapper[4646]: I1203 11:15:40.802250 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="proxy-httpd" containerID="cri-o://03e018aa4169175ce602de9e4767fff2300ce0f1dbc626845c717f2c9bde9c0c" gracePeriod=30 Dec 03 11:15:40 crc kubenswrapper[4646]: I1203 11:15:40.802248 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="ceilometer-notification-agent" containerID="cri-o://fe7acf9d5da8040edf3fcb08b5e2fbd194d97f1eda5ac4d210ccb34497e92fd9" gracePeriod=30 Dec 03 11:15:41 crc kubenswrapper[4646]: I1203 11:15:41.616124 4646 generic.go:334] "Generic (PLEG): container finished" podID="82837c9d-805c-43f1-90cf-16f4bd77c469" containerID="bd02bbc3ac3a06c5406acb8db6c4c07a7df6b6a2a98709fa794b2398e5e9dbd1" exitCode=143 Dec 03 11:15:41 crc kubenswrapper[4646]: I1203 11:15:41.616313 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"82837c9d-805c-43f1-90cf-16f4bd77c469","Type":"ContainerDied","Data":"bd02bbc3ac3a06c5406acb8db6c4c07a7df6b6a2a98709fa794b2398e5e9dbd1"} Dec 03 11:15:41 crc kubenswrapper[4646]: I1203 11:15:41.620000 4646 generic.go:334] "Generic (PLEG): container finished" podID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerID="03e018aa4169175ce602de9e4767fff2300ce0f1dbc626845c717f2c9bde9c0c" exitCode=0 Dec 03 11:15:41 crc kubenswrapper[4646]: I1203 11:15:41.620026 4646 generic.go:334] "Generic (PLEG): container finished" podID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerID="c0554e3a5c4c523da89bee4d069ea3535af17ebaf6da6238b7f045c1eec1ddf3" exitCode=2 Dec 03 11:15:41 crc kubenswrapper[4646]: I1203 11:15:41.620037 4646 generic.go:334] "Generic (PLEG): container finished" podID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerID="226337a866312357ba40f1a0db7ec0bf96a65f4eb46ed3994e41dd985a1ba1d2" exitCode=0 Dec 03 11:15:41 crc kubenswrapper[4646]: I1203 11:15:41.620101 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8","Type":"ContainerDied","Data":"03e018aa4169175ce602de9e4767fff2300ce0f1dbc626845c717f2c9bde9c0c"} Dec 03 11:15:41 crc kubenswrapper[4646]: I1203 11:15:41.620140 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8","Type":"ContainerDied","Data":"c0554e3a5c4c523da89bee4d069ea3535af17ebaf6da6238b7f045c1eec1ddf3"} Dec 03 11:15:41 crc kubenswrapper[4646]: I1203 11:15:41.620153 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8","Type":"ContainerDied","Data":"226337a866312357ba40f1a0db7ec0bf96a65f4eb46ed3994e41dd985a1ba1d2"} Dec 03 11:15:41 crc kubenswrapper[4646]: I1203 11:15:41.620402 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.475270 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.630986 4646 generic.go:334] "Generic (PLEG): container finished" podID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerID="fe7acf9d5da8040edf3fcb08b5e2fbd194d97f1eda5ac4d210ccb34497e92fd9" exitCode=0 Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.631973 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.632153 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8","Type":"ContainerDied","Data":"fe7acf9d5da8040edf3fcb08b5e2fbd194d97f1eda5ac4d210ccb34497e92fd9"} Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.632186 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8","Type":"ContainerDied","Data":"e67a8b26b0b0935ee3595921bb7aae50b42a1b9a59acf01321507fd827649d2b"} Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.632204 4646 scope.go:117] "RemoveContainer" containerID="03e018aa4169175ce602de9e4767fff2300ce0f1dbc626845c717f2c9bde9c0c" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.660325 4646 scope.go:117] "RemoveContainer" containerID="c0554e3a5c4c523da89bee4d069ea3535af17ebaf6da6238b7f045c1eec1ddf3" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.661132 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-run-httpd\") pod \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.661206 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-scripts\") pod \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.661386 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-config-data\") pod \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.661424 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-log-httpd\") pod \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.661456 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-ceilometer-tls-certs\") pod \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.661510 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-sg-core-conf-yaml\") pod \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.661565 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lzdp\" (UniqueName: \"kubernetes.io/projected/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-kube-api-access-2lzdp\") pod \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.661586 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-combined-ca-bundle\") pod \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\" (UID: \"1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8\") " Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.661762 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" (UID: "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.661992 4646 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.662551 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" (UID: "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.676410 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-kube-api-access-2lzdp" (OuterVolumeSpecName: "kube-api-access-2lzdp") pod "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" (UID: "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8"). InnerVolumeSpecName "kube-api-access-2lzdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.687150 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-scripts" (OuterVolumeSpecName: "scripts") pod "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" (UID: "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.688866 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" (UID: "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.708324 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" (UID: "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.757494 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" (UID: "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.764817 4646 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.764915 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lzdp\" (UniqueName: \"kubernetes.io/projected/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-kube-api-access-2lzdp\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.764980 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.765042 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.765098 4646 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.765161 4646 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.768874 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-config-data" (OuterVolumeSpecName: "config-data") pod "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" (UID: "1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.801686 4646 scope.go:117] "RemoveContainer" containerID="fe7acf9d5da8040edf3fcb08b5e2fbd194d97f1eda5ac4d210ccb34497e92fd9" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.827594 4646 scope.go:117] "RemoveContainer" containerID="226337a866312357ba40f1a0db7ec0bf96a65f4eb46ed3994e41dd985a1ba1d2" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.854647 4646 scope.go:117] "RemoveContainer" containerID="03e018aa4169175ce602de9e4767fff2300ce0f1dbc626845c717f2c9bde9c0c" Dec 03 11:15:42 crc kubenswrapper[4646]: E1203 11:15:42.855258 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03e018aa4169175ce602de9e4767fff2300ce0f1dbc626845c717f2c9bde9c0c\": container with ID starting with 03e018aa4169175ce602de9e4767fff2300ce0f1dbc626845c717f2c9bde9c0c not found: ID does not exist" containerID="03e018aa4169175ce602de9e4767fff2300ce0f1dbc626845c717f2c9bde9c0c" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.855315 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03e018aa4169175ce602de9e4767fff2300ce0f1dbc626845c717f2c9bde9c0c"} err="failed to get container status \"03e018aa4169175ce602de9e4767fff2300ce0f1dbc626845c717f2c9bde9c0c\": rpc error: code = NotFound desc = could not find container \"03e018aa4169175ce602de9e4767fff2300ce0f1dbc626845c717f2c9bde9c0c\": container with ID starting with 03e018aa4169175ce602de9e4767fff2300ce0f1dbc626845c717f2c9bde9c0c not found: ID does not exist" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.855368 4646 scope.go:117] "RemoveContainer" containerID="c0554e3a5c4c523da89bee4d069ea3535af17ebaf6da6238b7f045c1eec1ddf3" Dec 03 11:15:42 crc kubenswrapper[4646]: E1203 11:15:42.856062 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0554e3a5c4c523da89bee4d069ea3535af17ebaf6da6238b7f045c1eec1ddf3\": container with ID starting with c0554e3a5c4c523da89bee4d069ea3535af17ebaf6da6238b7f045c1eec1ddf3 not found: ID does not exist" containerID="c0554e3a5c4c523da89bee4d069ea3535af17ebaf6da6238b7f045c1eec1ddf3" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.856104 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0554e3a5c4c523da89bee4d069ea3535af17ebaf6da6238b7f045c1eec1ddf3"} err="failed to get container status \"c0554e3a5c4c523da89bee4d069ea3535af17ebaf6da6238b7f045c1eec1ddf3\": rpc error: code = NotFound desc = could not find container \"c0554e3a5c4c523da89bee4d069ea3535af17ebaf6da6238b7f045c1eec1ddf3\": container with ID starting with c0554e3a5c4c523da89bee4d069ea3535af17ebaf6da6238b7f045c1eec1ddf3 not found: ID does not exist" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.856128 4646 scope.go:117] "RemoveContainer" containerID="fe7acf9d5da8040edf3fcb08b5e2fbd194d97f1eda5ac4d210ccb34497e92fd9" Dec 03 11:15:42 crc kubenswrapper[4646]: E1203 11:15:42.856477 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe7acf9d5da8040edf3fcb08b5e2fbd194d97f1eda5ac4d210ccb34497e92fd9\": container with ID starting with fe7acf9d5da8040edf3fcb08b5e2fbd194d97f1eda5ac4d210ccb34497e92fd9 not found: ID does not exist" containerID="fe7acf9d5da8040edf3fcb08b5e2fbd194d97f1eda5ac4d210ccb34497e92fd9" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.856505 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe7acf9d5da8040edf3fcb08b5e2fbd194d97f1eda5ac4d210ccb34497e92fd9"} err="failed to get container status \"fe7acf9d5da8040edf3fcb08b5e2fbd194d97f1eda5ac4d210ccb34497e92fd9\": rpc error: code = NotFound desc = could not find container \"fe7acf9d5da8040edf3fcb08b5e2fbd194d97f1eda5ac4d210ccb34497e92fd9\": container with ID starting with fe7acf9d5da8040edf3fcb08b5e2fbd194d97f1eda5ac4d210ccb34497e92fd9 not found: ID does not exist" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.856526 4646 scope.go:117] "RemoveContainer" containerID="226337a866312357ba40f1a0db7ec0bf96a65f4eb46ed3994e41dd985a1ba1d2" Dec 03 11:15:42 crc kubenswrapper[4646]: E1203 11:15:42.856874 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"226337a866312357ba40f1a0db7ec0bf96a65f4eb46ed3994e41dd985a1ba1d2\": container with ID starting with 226337a866312357ba40f1a0db7ec0bf96a65f4eb46ed3994e41dd985a1ba1d2 not found: ID does not exist" containerID="226337a866312357ba40f1a0db7ec0bf96a65f4eb46ed3994e41dd985a1ba1d2" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.856923 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"226337a866312357ba40f1a0db7ec0bf96a65f4eb46ed3994e41dd985a1ba1d2"} err="failed to get container status \"226337a866312357ba40f1a0db7ec0bf96a65f4eb46ed3994e41dd985a1ba1d2\": rpc error: code = NotFound desc = could not find container \"226337a866312357ba40f1a0db7ec0bf96a65f4eb46ed3994e41dd985a1ba1d2\": container with ID starting with 226337a866312357ba40f1a0db7ec0bf96a65f4eb46ed3994e41dd985a1ba1d2 not found: ID does not exist" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.867385 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.969024 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:15:42 crc kubenswrapper[4646]: I1203 11:15:42.977905 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.000017 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:15:43 crc kubenswrapper[4646]: E1203 11:15:43.000436 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="ceilometer-notification-agent" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.000459 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="ceilometer-notification-agent" Dec 03 11:15:43 crc kubenswrapper[4646]: E1203 11:15:43.000478 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="sg-core" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.000487 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="sg-core" Dec 03 11:15:43 crc kubenswrapper[4646]: E1203 11:15:43.000507 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="proxy-httpd" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.000514 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="proxy-httpd" Dec 03 11:15:43 crc kubenswrapper[4646]: E1203 11:15:43.000541 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="ceilometer-central-agent" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.000548 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="ceilometer-central-agent" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.000720 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="ceilometer-central-agent" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.000739 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="sg-core" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.000749 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="proxy-httpd" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.000765 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" containerName="ceilometer-notification-agent" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.006269 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.014816 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.014995 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.015068 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.032982 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.172909 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/317f0cc8-e513-472a-86c4-9f49b831bbba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.173004 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wph66\" (UniqueName: \"kubernetes.io/projected/317f0cc8-e513-472a-86c4-9f49b831bbba-kube-api-access-wph66\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.173027 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/317f0cc8-e513-472a-86c4-9f49b831bbba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.173078 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/317f0cc8-e513-472a-86c4-9f49b831bbba-config-data\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.173098 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/317f0cc8-e513-472a-86c4-9f49b831bbba-scripts\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.173132 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317f0cc8-e513-472a-86c4-9f49b831bbba-run-httpd\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.173170 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/317f0cc8-e513-472a-86c4-9f49b831bbba-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.173253 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317f0cc8-e513-472a-86c4-9f49b831bbba-log-httpd\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.275400 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wph66\" (UniqueName: \"kubernetes.io/projected/317f0cc8-e513-472a-86c4-9f49b831bbba-kube-api-access-wph66\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.275446 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/317f0cc8-e513-472a-86c4-9f49b831bbba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.275489 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/317f0cc8-e513-472a-86c4-9f49b831bbba-config-data\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.275512 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/317f0cc8-e513-472a-86c4-9f49b831bbba-scripts\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.275533 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317f0cc8-e513-472a-86c4-9f49b831bbba-run-httpd\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.275572 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/317f0cc8-e513-472a-86c4-9f49b831bbba-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.275591 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317f0cc8-e513-472a-86c4-9f49b831bbba-log-httpd\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.275644 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/317f0cc8-e513-472a-86c4-9f49b831bbba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.276395 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317f0cc8-e513-472a-86c4-9f49b831bbba-log-httpd\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.276572 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317f0cc8-e513-472a-86c4-9f49b831bbba-run-httpd\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.279462 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/317f0cc8-e513-472a-86c4-9f49b831bbba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.279991 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/317f0cc8-e513-472a-86c4-9f49b831bbba-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.280134 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/317f0cc8-e513-472a-86c4-9f49b831bbba-scripts\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.290095 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/317f0cc8-e513-472a-86c4-9f49b831bbba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.292489 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/317f0cc8-e513-472a-86c4-9f49b831bbba-config-data\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.294629 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wph66\" (UniqueName: \"kubernetes.io/projected/317f0cc8-e513-472a-86c4-9f49b831bbba-kube-api-access-wph66\") pod \"ceilometer-0\" (UID: \"317f0cc8-e513-472a-86c4-9f49b831bbba\") " pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.369644 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.824869 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 11:15:43 crc kubenswrapper[4646]: W1203 11:15:43.840112 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod317f0cc8_e513_472a_86c4_9f49b831bbba.slice/crio-ffc16c4bba0a8b3b97f30cbc97254b686504c686f72ad1262653275a002c3fc0 WatchSource:0}: Error finding container ffc16c4bba0a8b3b97f30cbc97254b686504c686f72ad1262653275a002c3fc0: Status 404 returned error can't find the container with id ffc16c4bba0a8b3b97f30cbc97254b686504c686f72ad1262653275a002c3fc0 Dec 03 11:15:43 crc kubenswrapper[4646]: I1203 11:15:43.868678 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8" path="/var/lib/kubelet/pods/1bfbebd2-1cc8-4a63-96cc-0328cea1d9a8/volumes" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.247284 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.414445 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82837c9d-805c-43f1-90cf-16f4bd77c469-logs\") pod \"82837c9d-805c-43f1-90cf-16f4bd77c469\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.414758 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82837c9d-805c-43f1-90cf-16f4bd77c469-combined-ca-bundle\") pod \"82837c9d-805c-43f1-90cf-16f4bd77c469\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.415001 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82837c9d-805c-43f1-90cf-16f4bd77c469-logs" (OuterVolumeSpecName: "logs") pod "82837c9d-805c-43f1-90cf-16f4bd77c469" (UID: "82837c9d-805c-43f1-90cf-16f4bd77c469"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.415259 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prbcm\" (UniqueName: \"kubernetes.io/projected/82837c9d-805c-43f1-90cf-16f4bd77c469-kube-api-access-prbcm\") pod \"82837c9d-805c-43f1-90cf-16f4bd77c469\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.415355 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82837c9d-805c-43f1-90cf-16f4bd77c469-config-data\") pod \"82837c9d-805c-43f1-90cf-16f4bd77c469\" (UID: \"82837c9d-805c-43f1-90cf-16f4bd77c469\") " Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.415794 4646 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82837c9d-805c-43f1-90cf-16f4bd77c469-logs\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.447580 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82837c9d-805c-43f1-90cf-16f4bd77c469-config-data" (OuterVolumeSpecName: "config-data") pod "82837c9d-805c-43f1-90cf-16f4bd77c469" (UID: "82837c9d-805c-43f1-90cf-16f4bd77c469"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.449653 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82837c9d-805c-43f1-90cf-16f4bd77c469-kube-api-access-prbcm" (OuterVolumeSpecName: "kube-api-access-prbcm") pod "82837c9d-805c-43f1-90cf-16f4bd77c469" (UID: "82837c9d-805c-43f1-90cf-16f4bd77c469"). InnerVolumeSpecName "kube-api-access-prbcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.490945 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82837c9d-805c-43f1-90cf-16f4bd77c469-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82837c9d-805c-43f1-90cf-16f4bd77c469" (UID: "82837c9d-805c-43f1-90cf-16f4bd77c469"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.516788 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82837c9d-805c-43f1-90cf-16f4bd77c469-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.516899 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prbcm\" (UniqueName: \"kubernetes.io/projected/82837c9d-805c-43f1-90cf-16f4bd77c469-kube-api-access-prbcm\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.516958 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82837c9d-805c-43f1-90cf-16f4bd77c469-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.655499 4646 generic.go:334] "Generic (PLEG): container finished" podID="82837c9d-805c-43f1-90cf-16f4bd77c469" containerID="c44b0a9d8d679da495cfad1c2f86c0ec38e16d3ad75edd95b20c8a3b7bc5513c" exitCode=0 Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.655954 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.658285 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"82837c9d-805c-43f1-90cf-16f4bd77c469","Type":"ContainerDied","Data":"c44b0a9d8d679da495cfad1c2f86c0ec38e16d3ad75edd95b20c8a3b7bc5513c"} Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.658367 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"82837c9d-805c-43f1-90cf-16f4bd77c469","Type":"ContainerDied","Data":"12efbc22dbacf165d04839bb1ce1503b00185dd11f965876ff7bba2526edab20"} Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.658397 4646 scope.go:117] "RemoveContainer" containerID="c44b0a9d8d679da495cfad1c2f86c0ec38e16d3ad75edd95b20c8a3b7bc5513c" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.659580 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"317f0cc8-e513-472a-86c4-9f49b831bbba","Type":"ContainerStarted","Data":"ffc16c4bba0a8b3b97f30cbc97254b686504c686f72ad1262653275a002c3fc0"} Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.715452 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.717225 4646 scope.go:117] "RemoveContainer" containerID="bd02bbc3ac3a06c5406acb8db6c4c07a7df6b6a2a98709fa794b2398e5e9dbd1" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.738427 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.754797 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:44 crc kubenswrapper[4646]: E1203 11:15:44.755226 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82837c9d-805c-43f1-90cf-16f4bd77c469" containerName="nova-api-api" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.755248 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="82837c9d-805c-43f1-90cf-16f4bd77c469" containerName="nova-api-api" Dec 03 11:15:44 crc kubenswrapper[4646]: E1203 11:15:44.755289 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82837c9d-805c-43f1-90cf-16f4bd77c469" containerName="nova-api-log" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.755296 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="82837c9d-805c-43f1-90cf-16f4bd77c469" containerName="nova-api-log" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.755483 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="82837c9d-805c-43f1-90cf-16f4bd77c469" containerName="nova-api-api" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.755499 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="82837c9d-805c-43f1-90cf-16f4bd77c469" containerName="nova-api-log" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.757214 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.759825 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.760014 4646 scope.go:117] "RemoveContainer" containerID="c44b0a9d8d679da495cfad1c2f86c0ec38e16d3ad75edd95b20c8a3b7bc5513c" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.761286 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.767359 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 11:15:44 crc kubenswrapper[4646]: E1203 11:15:44.767371 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c44b0a9d8d679da495cfad1c2f86c0ec38e16d3ad75edd95b20c8a3b7bc5513c\": container with ID starting with c44b0a9d8d679da495cfad1c2f86c0ec38e16d3ad75edd95b20c8a3b7bc5513c not found: ID does not exist" containerID="c44b0a9d8d679da495cfad1c2f86c0ec38e16d3ad75edd95b20c8a3b7bc5513c" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.767402 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c44b0a9d8d679da495cfad1c2f86c0ec38e16d3ad75edd95b20c8a3b7bc5513c"} err="failed to get container status \"c44b0a9d8d679da495cfad1c2f86c0ec38e16d3ad75edd95b20c8a3b7bc5513c\": rpc error: code = NotFound desc = could not find container \"c44b0a9d8d679da495cfad1c2f86c0ec38e16d3ad75edd95b20c8a3b7bc5513c\": container with ID starting with c44b0a9d8d679da495cfad1c2f86c0ec38e16d3ad75edd95b20c8a3b7bc5513c not found: ID does not exist" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.767426 4646 scope.go:117] "RemoveContainer" containerID="bd02bbc3ac3a06c5406acb8db6c4c07a7df6b6a2a98709fa794b2398e5e9dbd1" Dec 03 11:15:44 crc kubenswrapper[4646]: E1203 11:15:44.769387 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd02bbc3ac3a06c5406acb8db6c4c07a7df6b6a2a98709fa794b2398e5e9dbd1\": container with ID starting with bd02bbc3ac3a06c5406acb8db6c4c07a7df6b6a2a98709fa794b2398e5e9dbd1 not found: ID does not exist" containerID="bd02bbc3ac3a06c5406acb8db6c4c07a7df6b6a2a98709fa794b2398e5e9dbd1" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.769409 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd02bbc3ac3a06c5406acb8db6c4c07a7df6b6a2a98709fa794b2398e5e9dbd1"} err="failed to get container status \"bd02bbc3ac3a06c5406acb8db6c4c07a7df6b6a2a98709fa794b2398e5e9dbd1\": rpc error: code = NotFound desc = could not find container \"bd02bbc3ac3a06c5406acb8db6c4c07a7df6b6a2a98709fa794b2398e5e9dbd1\": container with ID starting with bd02bbc3ac3a06c5406acb8db6c4c07a7df6b6a2a98709fa794b2398e5e9dbd1 not found: ID does not exist" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.775929 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.825987 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.826027 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-config-data\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.826096 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a757a9cb-1675-44f9-ba8e-ac51910c25b5-logs\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.826194 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-public-tls-certs\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.826251 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.826356 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkgwg\" (UniqueName: \"kubernetes.io/projected/a757a9cb-1675-44f9-ba8e-ac51910c25b5-kube-api-access-lkgwg\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: E1203 11:15:44.826863 4646 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82837c9d_805c_43f1_90cf_16f4bd77c469.slice/crio-12efbc22dbacf165d04839bb1ce1503b00185dd11f965876ff7bba2526edab20\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82837c9d_805c_43f1_90cf_16f4bd77c469.slice\": RecentStats: unable to find data in memory cache]" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.928036 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a757a9cb-1675-44f9-ba8e-ac51910c25b5-logs\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.928452 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a757a9cb-1675-44f9-ba8e-ac51910c25b5-logs\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.928579 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-public-tls-certs\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.928599 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.928654 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkgwg\" (UniqueName: \"kubernetes.io/projected/a757a9cb-1675-44f9-ba8e-ac51910c25b5-kube-api-access-lkgwg\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.928714 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.928732 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-config-data\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.935077 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-public-tls-certs\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.935371 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-config-data\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.935833 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.943837 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:44 crc kubenswrapper[4646]: I1203 11:15:44.950864 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkgwg\" (UniqueName: \"kubernetes.io/projected/a757a9cb-1675-44f9-ba8e-ac51910c25b5-kube-api-access-lkgwg\") pod \"nova-api-0\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " pod="openstack/nova-api-0" Dec 03 11:15:45 crc kubenswrapper[4646]: I1203 11:15:45.108137 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 11:15:45 crc kubenswrapper[4646]: I1203 11:15:45.259569 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:45 crc kubenswrapper[4646]: I1203 11:15:45.297693 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:45 crc kubenswrapper[4646]: I1203 11:15:45.677042 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:45 crc kubenswrapper[4646]: I1203 11:15:45.691012 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"317f0cc8-e513-472a-86c4-9f49b831bbba","Type":"ContainerStarted","Data":"043278feb46a8e44f70baacb4a1f7898890e757e2e2f8f2f3cd6a65adc7e0da5"} Dec 03 11:15:45 crc kubenswrapper[4646]: W1203 11:15:45.749357 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda757a9cb_1675_44f9_ba8e_ac51910c25b5.slice/crio-209d588a499e3c403b96ead52c66f1c0aa98322e2e4e00dc6162443e76fca361 WatchSource:0}: Error finding container 209d588a499e3c403b96ead52c66f1c0aa98322e2e4e00dc6162443e76fca361: Status 404 returned error can't find the container with id 209d588a499e3c403b96ead52c66f1c0aa98322e2e4e00dc6162443e76fca361 Dec 03 11:15:45 crc kubenswrapper[4646]: I1203 11:15:45.845451 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 03 11:15:45 crc kubenswrapper[4646]: I1203 11:15:45.890022 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82837c9d-805c-43f1-90cf-16f4bd77c469" path="/var/lib/kubelet/pods/82837c9d-805c-43f1-90cf-16f4bd77c469/volumes" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.153400 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-gckcg"] Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.155288 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.158488 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.158723 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.162262 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-gckcg"] Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.163364 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gckcg\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.164657 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-config-data\") pod \"nova-cell1-cell-mapping-gckcg\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.165004 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-scripts\") pod \"nova-cell1-cell-mapping-gckcg\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.165110 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7thtt\" (UniqueName: \"kubernetes.io/projected/c171132a-33a9-4df3-8bb3-1789bb031da5-kube-api-access-7thtt\") pod \"nova-cell1-cell-mapping-gckcg\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.267216 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7thtt\" (UniqueName: \"kubernetes.io/projected/c171132a-33a9-4df3-8bb3-1789bb031da5-kube-api-access-7thtt\") pod \"nova-cell1-cell-mapping-gckcg\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.267271 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gckcg\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.267312 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-config-data\") pod \"nova-cell1-cell-mapping-gckcg\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.268292 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-scripts\") pod \"nova-cell1-cell-mapping-gckcg\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.271647 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-scripts\") pod \"nova-cell1-cell-mapping-gckcg\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.272384 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-config-data\") pod \"nova-cell1-cell-mapping-gckcg\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.272482 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gckcg\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.295889 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7thtt\" (UniqueName: \"kubernetes.io/projected/c171132a-33a9-4df3-8bb3-1789bb031da5-kube-api-access-7thtt\") pod \"nova-cell1-cell-mapping-gckcg\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.558096 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.723908 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a757a9cb-1675-44f9-ba8e-ac51910c25b5","Type":"ContainerStarted","Data":"7a43a3cd0da98f21dbcc65af77ea8a95ee3eceacf3597a9b21fa6d522d5ee723"} Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.724228 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a757a9cb-1675-44f9-ba8e-ac51910c25b5","Type":"ContainerStarted","Data":"9f9d7ed87925fd43fa7fd68bcb166d0b07b4a62b8122142d78cd3b03b0084248"} Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.724243 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a757a9cb-1675-44f9-ba8e-ac51910c25b5","Type":"ContainerStarted","Data":"209d588a499e3c403b96ead52c66f1c0aa98322e2e4e00dc6162443e76fca361"} Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.763987 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.763967224 podStartE2EDuration="2.763967224s" podCreationTimestamp="2025-12-03 11:15:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:15:46.748868011 +0000 UTC m=+1323.211924156" watchObservedRunningTime="2025-12-03 11:15:46.763967224 +0000 UTC m=+1323.227023359" Dec 03 11:15:46 crc kubenswrapper[4646]: I1203 11:15:46.912068 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-gckcg"] Dec 03 11:15:47 crc kubenswrapper[4646]: I1203 11:15:47.732054 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gckcg" event={"ID":"c171132a-33a9-4df3-8bb3-1789bb031da5","Type":"ContainerStarted","Data":"7d185132f8aab22aff0d4aa5add5f3a1da579b91cebeb0b4e1f10046c1b01d55"} Dec 03 11:15:47 crc kubenswrapper[4646]: I1203 11:15:47.732598 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gckcg" event={"ID":"c171132a-33a9-4df3-8bb3-1789bb031da5","Type":"ContainerStarted","Data":"db7488470d8754ef47c6454e80beff979d62d9e11f3c427df3f1a043107c8e12"} Dec 03 11:15:47 crc kubenswrapper[4646]: I1203 11:15:47.734989 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"317f0cc8-e513-472a-86c4-9f49b831bbba","Type":"ContainerStarted","Data":"e99a4f76d807e2a9df2d340fbf552c611373082ae3bd9dd13ed5c41a790865c2"} Dec 03 11:15:47 crc kubenswrapper[4646]: I1203 11:15:47.735034 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"317f0cc8-e513-472a-86c4-9f49b831bbba","Type":"ContainerStarted","Data":"cd77c82ba4e071cfd33cfed6eee4064ff9db9196e9c0f9604bf2b082fc7ccf4f"} Dec 03 11:15:47 crc kubenswrapper[4646]: I1203 11:15:47.748321 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-gckcg" podStartSLOduration=1.748300194 podStartE2EDuration="1.748300194s" podCreationTimestamp="2025-12-03 11:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:15:47.747515652 +0000 UTC m=+1324.210571787" watchObservedRunningTime="2025-12-03 11:15:47.748300194 +0000 UTC m=+1324.211356329" Dec 03 11:15:48 crc kubenswrapper[4646]: I1203 11:15:48.387121 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:15:48 crc kubenswrapper[4646]: I1203 11:15:48.450888 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-d6p7n"] Dec 03 11:15:48 crc kubenswrapper[4646]: I1203 11:15:48.451118 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" podUID="16b0f468-ba7c-4c95-bac0-9a19912975c2" containerName="dnsmasq-dns" containerID="cri-o://7f221a9351df0f408a440796ef0676ff07011535f8fd18dd0ced5ee80ec5b867" gracePeriod=10 Dec 03 11:15:48 crc kubenswrapper[4646]: I1203 11:15:48.770073 4646 generic.go:334] "Generic (PLEG): container finished" podID="16b0f468-ba7c-4c95-bac0-9a19912975c2" containerID="7f221a9351df0f408a440796ef0676ff07011535f8fd18dd0ced5ee80ec5b867" exitCode=0 Dec 03 11:15:48 crc kubenswrapper[4646]: I1203 11:15:48.771124 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" event={"ID":"16b0f468-ba7c-4c95-bac0-9a19912975c2","Type":"ContainerDied","Data":"7f221a9351df0f408a440796ef0676ff07011535f8fd18dd0ced5ee80ec5b867"} Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.272720 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.352431 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-config\") pod \"16b0f468-ba7c-4c95-bac0-9a19912975c2\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.352608 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-ovsdbserver-nb\") pod \"16b0f468-ba7c-4c95-bac0-9a19912975c2\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.352636 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbdvt\" (UniqueName: \"kubernetes.io/projected/16b0f468-ba7c-4c95-bac0-9a19912975c2-kube-api-access-zbdvt\") pod \"16b0f468-ba7c-4c95-bac0-9a19912975c2\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.352678 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-dns-svc\") pod \"16b0f468-ba7c-4c95-bac0-9a19912975c2\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.352822 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-ovsdbserver-sb\") pod \"16b0f468-ba7c-4c95-bac0-9a19912975c2\" (UID: \"16b0f468-ba7c-4c95-bac0-9a19912975c2\") " Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.356512 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16b0f468-ba7c-4c95-bac0-9a19912975c2-kube-api-access-zbdvt" (OuterVolumeSpecName: "kube-api-access-zbdvt") pod "16b0f468-ba7c-4c95-bac0-9a19912975c2" (UID: "16b0f468-ba7c-4c95-bac0-9a19912975c2"). InnerVolumeSpecName "kube-api-access-zbdvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.408924 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "16b0f468-ba7c-4c95-bac0-9a19912975c2" (UID: "16b0f468-ba7c-4c95-bac0-9a19912975c2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.422144 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "16b0f468-ba7c-4c95-bac0-9a19912975c2" (UID: "16b0f468-ba7c-4c95-bac0-9a19912975c2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.437620 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "16b0f468-ba7c-4c95-bac0-9a19912975c2" (UID: "16b0f468-ba7c-4c95-bac0-9a19912975c2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.438222 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-config" (OuterVolumeSpecName: "config") pod "16b0f468-ba7c-4c95-bac0-9a19912975c2" (UID: "16b0f468-ba7c-4c95-bac0-9a19912975c2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.454732 4646 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.454773 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.454785 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.454793 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16b0f468-ba7c-4c95-bac0-9a19912975c2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.454802 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbdvt\" (UniqueName: \"kubernetes.io/projected/16b0f468-ba7c-4c95-bac0-9a19912975c2-kube-api-access-zbdvt\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.789163 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"317f0cc8-e513-472a-86c4-9f49b831bbba","Type":"ContainerStarted","Data":"a6e2dee5dd669c79268fff2825d50a834bf8337987c041d01f412804de0d89c1"} Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.789301 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.797100 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" event={"ID":"16b0f468-ba7c-4c95-bac0-9a19912975c2","Type":"ContainerDied","Data":"4f715dd0c7f30e0f5688f620d8e81fcd939543930eef95db9c5436c7038f3c50"} Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.797154 4646 scope.go:117] "RemoveContainer" containerID="7f221a9351df0f408a440796ef0676ff07011535f8fd18dd0ced5ee80ec5b867" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.797562 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-d6p7n" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.833195 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.685287956 podStartE2EDuration="7.833177355s" podCreationTimestamp="2025-12-03 11:15:42 +0000 UTC" firstStartedPulling="2025-12-03 11:15:43.842250693 +0000 UTC m=+1320.305306828" lastFinishedPulling="2025-12-03 11:15:48.990140092 +0000 UTC m=+1325.453196227" observedRunningTime="2025-12-03 11:15:49.825925762 +0000 UTC m=+1326.288981897" watchObservedRunningTime="2025-12-03 11:15:49.833177355 +0000 UTC m=+1326.296233490" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.862070 4646 scope.go:117] "RemoveContainer" containerID="32660307339de85a58f3445c2402560e1088d71134c3e721fcf99c7b1db099b0" Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.878554 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-d6p7n"] Dec 03 11:15:49 crc kubenswrapper[4646]: I1203 11:15:49.883937 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-d6p7n"] Dec 03 11:15:51 crc kubenswrapper[4646]: I1203 11:15:51.858735 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16b0f468-ba7c-4c95-bac0-9a19912975c2" path="/var/lib/kubelet/pods/16b0f468-ba7c-4c95-bac0-9a19912975c2/volumes" Dec 03 11:15:52 crc kubenswrapper[4646]: I1203 11:15:52.831679 4646 generic.go:334] "Generic (PLEG): container finished" podID="c171132a-33a9-4df3-8bb3-1789bb031da5" containerID="7d185132f8aab22aff0d4aa5add5f3a1da579b91cebeb0b4e1f10046c1b01d55" exitCode=0 Dec 03 11:15:52 crc kubenswrapper[4646]: I1203 11:15:52.831783 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gckcg" event={"ID":"c171132a-33a9-4df3-8bb3-1789bb031da5","Type":"ContainerDied","Data":"7d185132f8aab22aff0d4aa5add5f3a1da579b91cebeb0b4e1f10046c1b01d55"} Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.227911 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.273194 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-config-data\") pod \"c171132a-33a9-4df3-8bb3-1789bb031da5\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.273574 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-scripts\") pod \"c171132a-33a9-4df3-8bb3-1789bb031da5\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.273620 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7thtt\" (UniqueName: \"kubernetes.io/projected/c171132a-33a9-4df3-8bb3-1789bb031da5-kube-api-access-7thtt\") pod \"c171132a-33a9-4df3-8bb3-1789bb031da5\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.273709 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-combined-ca-bundle\") pod \"c171132a-33a9-4df3-8bb3-1789bb031da5\" (UID: \"c171132a-33a9-4df3-8bb3-1789bb031da5\") " Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.287945 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-scripts" (OuterVolumeSpecName: "scripts") pod "c171132a-33a9-4df3-8bb3-1789bb031da5" (UID: "c171132a-33a9-4df3-8bb3-1789bb031da5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.288357 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c171132a-33a9-4df3-8bb3-1789bb031da5-kube-api-access-7thtt" (OuterVolumeSpecName: "kube-api-access-7thtt") pod "c171132a-33a9-4df3-8bb3-1789bb031da5" (UID: "c171132a-33a9-4df3-8bb3-1789bb031da5"). InnerVolumeSpecName "kube-api-access-7thtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.303479 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c171132a-33a9-4df3-8bb3-1789bb031da5" (UID: "c171132a-33a9-4df3-8bb3-1789bb031da5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.308891 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-config-data" (OuterVolumeSpecName: "config-data") pod "c171132a-33a9-4df3-8bb3-1789bb031da5" (UID: "c171132a-33a9-4df3-8bb3-1789bb031da5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.375369 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.375403 4646 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.375414 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7thtt\" (UniqueName: \"kubernetes.io/projected/c171132a-33a9-4df3-8bb3-1789bb031da5-kube-api-access-7thtt\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.375426 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c171132a-33a9-4df3-8bb3-1789bb031da5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.852303 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gckcg" event={"ID":"c171132a-33a9-4df3-8bb3-1789bb031da5","Type":"ContainerDied","Data":"db7488470d8754ef47c6454e80beff979d62d9e11f3c427df3f1a043107c8e12"} Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.852649 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db7488470d8754ef47c6454e80beff979d62d9e11f3c427df3f1a043107c8e12" Dec 03 11:15:54 crc kubenswrapper[4646]: I1203 11:15:54.852364 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gckcg" Dec 03 11:15:55 crc kubenswrapper[4646]: I1203 11:15:55.029883 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:55 crc kubenswrapper[4646]: I1203 11:15:55.030856 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a757a9cb-1675-44f9-ba8e-ac51910c25b5" containerName="nova-api-api" containerID="cri-o://7a43a3cd0da98f21dbcc65af77ea8a95ee3eceacf3597a9b21fa6d522d5ee723" gracePeriod=30 Dec 03 11:15:55 crc kubenswrapper[4646]: I1203 11:15:55.031193 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a757a9cb-1675-44f9-ba8e-ac51910c25b5" containerName="nova-api-log" containerID="cri-o://9f9d7ed87925fd43fa7fd68bcb166d0b07b4a62b8122142d78cd3b03b0084248" gracePeriod=30 Dec 03 11:15:55 crc kubenswrapper[4646]: I1203 11:15:55.052141 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 11:15:55 crc kubenswrapper[4646]: I1203 11:15:55.052374 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a7a75658-d6ae-4875-a787-0b3e49697499" containerName="nova-scheduler-scheduler" containerID="cri-o://5265a31fb663da564a289470d89f019e099cf570cad86d4d793f454e0ede77fa" gracePeriod=30 Dec 03 11:15:55 crc kubenswrapper[4646]: I1203 11:15:55.098417 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:55 crc kubenswrapper[4646]: I1203 11:15:55.098695 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e5dc2d79-d598-4b4d-80fd-4ac11612b32f" containerName="nova-metadata-log" containerID="cri-o://5e09820f691f7079067c490be8272a91f0bb4eec739e409bb9604ab4d889a4b4" gracePeriod=30 Dec 03 11:15:55 crc kubenswrapper[4646]: I1203 11:15:55.099061 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e5dc2d79-d598-4b4d-80fd-4ac11612b32f" containerName="nova-metadata-metadata" containerID="cri-o://1b476e700d6473b5090594015176b8ef1b20eab1116d439a30d346cb91ea20d6" gracePeriod=30 Dec 03 11:15:55 crc kubenswrapper[4646]: E1203 11:15:55.114610 4646 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc171132a_33a9_4df3_8bb3_1789bb031da5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc171132a_33a9_4df3_8bb3_1789bb031da5.slice/crio-db7488470d8754ef47c6454e80beff979d62d9e11f3c427df3f1a043107c8e12\": RecentStats: unable to find data in memory cache]" Dec 03 11:15:55 crc kubenswrapper[4646]: I1203 11:15:55.884372 4646 generic.go:334] "Generic (PLEG): container finished" podID="a757a9cb-1675-44f9-ba8e-ac51910c25b5" containerID="7a43a3cd0da98f21dbcc65af77ea8a95ee3eceacf3597a9b21fa6d522d5ee723" exitCode=0 Dec 03 11:15:55 crc kubenswrapper[4646]: I1203 11:15:55.884955 4646 generic.go:334] "Generic (PLEG): container finished" podID="a757a9cb-1675-44f9-ba8e-ac51910c25b5" containerID="9f9d7ed87925fd43fa7fd68bcb166d0b07b4a62b8122142d78cd3b03b0084248" exitCode=143 Dec 03 11:15:55 crc kubenswrapper[4646]: I1203 11:15:55.884456 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a757a9cb-1675-44f9-ba8e-ac51910c25b5","Type":"ContainerDied","Data":"7a43a3cd0da98f21dbcc65af77ea8a95ee3eceacf3597a9b21fa6d522d5ee723"} Dec 03 11:15:55 crc kubenswrapper[4646]: I1203 11:15:55.885085 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a757a9cb-1675-44f9-ba8e-ac51910c25b5","Type":"ContainerDied","Data":"9f9d7ed87925fd43fa7fd68bcb166d0b07b4a62b8122142d78cd3b03b0084248"} Dec 03 11:15:55 crc kubenswrapper[4646]: I1203 11:15:55.890964 4646 generic.go:334] "Generic (PLEG): container finished" podID="e5dc2d79-d598-4b4d-80fd-4ac11612b32f" containerID="5e09820f691f7079067c490be8272a91f0bb4eec739e409bb9604ab4d889a4b4" exitCode=143 Dec 03 11:15:55 crc kubenswrapper[4646]: I1203 11:15:55.891002 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5dc2d79-d598-4b4d-80fd-4ac11612b32f","Type":"ContainerDied","Data":"5e09820f691f7079067c490be8272a91f0bb4eec739e409bb9604ab4d889a4b4"} Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.055032 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.117968 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-internal-tls-certs\") pod \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.118123 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-combined-ca-bundle\") pod \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.118206 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-config-data\") pod \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.118232 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a757a9cb-1675-44f9-ba8e-ac51910c25b5-logs\") pod \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.118250 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-public-tls-certs\") pod \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.118285 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkgwg\" (UniqueName: \"kubernetes.io/projected/a757a9cb-1675-44f9-ba8e-ac51910c25b5-kube-api-access-lkgwg\") pod \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\" (UID: \"a757a9cb-1675-44f9-ba8e-ac51910c25b5\") " Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.119043 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a757a9cb-1675-44f9-ba8e-ac51910c25b5-logs" (OuterVolumeSpecName: "logs") pod "a757a9cb-1675-44f9-ba8e-ac51910c25b5" (UID: "a757a9cb-1675-44f9-ba8e-ac51910c25b5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.119414 4646 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a757a9cb-1675-44f9-ba8e-ac51910c25b5-logs\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.123645 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a757a9cb-1675-44f9-ba8e-ac51910c25b5-kube-api-access-lkgwg" (OuterVolumeSpecName: "kube-api-access-lkgwg") pod "a757a9cb-1675-44f9-ba8e-ac51910c25b5" (UID: "a757a9cb-1675-44f9-ba8e-ac51910c25b5"). InnerVolumeSpecName "kube-api-access-lkgwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.158530 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-config-data" (OuterVolumeSpecName: "config-data") pod "a757a9cb-1675-44f9-ba8e-ac51910c25b5" (UID: "a757a9cb-1675-44f9-ba8e-ac51910c25b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.158562 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a757a9cb-1675-44f9-ba8e-ac51910c25b5" (UID: "a757a9cb-1675-44f9-ba8e-ac51910c25b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.178520 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a757a9cb-1675-44f9-ba8e-ac51910c25b5" (UID: "a757a9cb-1675-44f9-ba8e-ac51910c25b5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.192078 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a757a9cb-1675-44f9-ba8e-ac51910c25b5" (UID: "a757a9cb-1675-44f9-ba8e-ac51910c25b5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.221522 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.221553 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.221563 4646 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.221573 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkgwg\" (UniqueName: \"kubernetes.io/projected/a757a9cb-1675-44f9-ba8e-ac51910c25b5-kube-api-access-lkgwg\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.221583 4646 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a757a9cb-1675-44f9-ba8e-ac51910c25b5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.899418 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a757a9cb-1675-44f9-ba8e-ac51910c25b5","Type":"ContainerDied","Data":"209d588a499e3c403b96ead52c66f1c0aa98322e2e4e00dc6162443e76fca361"} Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.899502 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.900407 4646 scope.go:117] "RemoveContainer" containerID="7a43a3cd0da98f21dbcc65af77ea8a95ee3eceacf3597a9b21fa6d522d5ee723" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.927483 4646 scope.go:117] "RemoveContainer" containerID="9f9d7ed87925fd43fa7fd68bcb166d0b07b4a62b8122142d78cd3b03b0084248" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.939654 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.949455 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.962388 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:56 crc kubenswrapper[4646]: E1203 11:15:56.963072 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16b0f468-ba7c-4c95-bac0-9a19912975c2" containerName="init" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.963145 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="16b0f468-ba7c-4c95-bac0-9a19912975c2" containerName="init" Dec 03 11:15:56 crc kubenswrapper[4646]: E1203 11:15:56.963214 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a757a9cb-1675-44f9-ba8e-ac51910c25b5" containerName="nova-api-log" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.963264 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="a757a9cb-1675-44f9-ba8e-ac51910c25b5" containerName="nova-api-log" Dec 03 11:15:56 crc kubenswrapper[4646]: E1203 11:15:56.963322 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16b0f468-ba7c-4c95-bac0-9a19912975c2" containerName="dnsmasq-dns" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.963414 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="16b0f468-ba7c-4c95-bac0-9a19912975c2" containerName="dnsmasq-dns" Dec 03 11:15:56 crc kubenswrapper[4646]: E1203 11:15:56.963485 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c171132a-33a9-4df3-8bb3-1789bb031da5" containerName="nova-manage" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.963539 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="c171132a-33a9-4df3-8bb3-1789bb031da5" containerName="nova-manage" Dec 03 11:15:56 crc kubenswrapper[4646]: E1203 11:15:56.963600 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a757a9cb-1675-44f9-ba8e-ac51910c25b5" containerName="nova-api-api" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.963680 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="a757a9cb-1675-44f9-ba8e-ac51910c25b5" containerName="nova-api-api" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.963911 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="16b0f468-ba7c-4c95-bac0-9a19912975c2" containerName="dnsmasq-dns" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.963980 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="a757a9cb-1675-44f9-ba8e-ac51910c25b5" containerName="nova-api-log" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.964040 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="a757a9cb-1675-44f9-ba8e-ac51910c25b5" containerName="nova-api-api" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.964112 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="c171132a-33a9-4df3-8bb3-1789bb031da5" containerName="nova-manage" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.965056 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.969876 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.970281 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.972281 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 11:15:56 crc kubenswrapper[4646]: I1203 11:15:56.973622 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.034877 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c9b16d-acd9-453f-adb9-505992d25ec2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.034935 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99c9b16d-acd9-453f-adb9-505992d25ec2-logs\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.034951 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c9b16d-acd9-453f-adb9-505992d25ec2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.034984 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c9b16d-acd9-453f-adb9-505992d25ec2-config-data\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.035087 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6nxq\" (UniqueName: \"kubernetes.io/projected/99c9b16d-acd9-453f-adb9-505992d25ec2-kube-api-access-w6nxq\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.035120 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c9b16d-acd9-453f-adb9-505992d25ec2-public-tls-certs\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.136200 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6nxq\" (UniqueName: \"kubernetes.io/projected/99c9b16d-acd9-453f-adb9-505992d25ec2-kube-api-access-w6nxq\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.136466 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c9b16d-acd9-453f-adb9-505992d25ec2-public-tls-certs\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.136703 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c9b16d-acd9-453f-adb9-505992d25ec2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.136944 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99c9b16d-acd9-453f-adb9-505992d25ec2-logs\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.137050 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c9b16d-acd9-453f-adb9-505992d25ec2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.137160 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c9b16d-acd9-453f-adb9-505992d25ec2-config-data\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.137699 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99c9b16d-acd9-453f-adb9-505992d25ec2-logs\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.142946 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c9b16d-acd9-453f-adb9-505992d25ec2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.143043 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c9b16d-acd9-453f-adb9-505992d25ec2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.143223 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c9b16d-acd9-453f-adb9-505992d25ec2-config-data\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.143422 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c9b16d-acd9-453f-adb9-505992d25ec2-public-tls-certs\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.152503 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6nxq\" (UniqueName: \"kubernetes.io/projected/99c9b16d-acd9-453f-adb9-505992d25ec2-kube-api-access-w6nxq\") pod \"nova-api-0\" (UID: \"99c9b16d-acd9-453f-adb9-505992d25ec2\") " pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.281903 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.708359 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.864819 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a757a9cb-1675-44f9-ba8e-ac51910c25b5" path="/var/lib/kubelet/pods/a757a9cb-1675-44f9-ba8e-ac51910c25b5/volumes" Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.911999 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"99c9b16d-acd9-453f-adb9-505992d25ec2","Type":"ContainerStarted","Data":"4e3fdfd431676cfbae47fbc9fad40dc2f11af8cd034df8018266372edfe128be"} Dec 03 11:15:57 crc kubenswrapper[4646]: I1203 11:15:57.912350 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"99c9b16d-acd9-453f-adb9-505992d25ec2","Type":"ContainerStarted","Data":"42b476d365addd8887d7e75ba937d9fe59b764708b9242038f7d795d89d7a5db"} Dec 03 11:15:58 crc kubenswrapper[4646]: I1203 11:15:58.895657 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 11:15:58 crc kubenswrapper[4646]: I1203 11:15:58.931806 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"99c9b16d-acd9-453f-adb9-505992d25ec2","Type":"ContainerStarted","Data":"957e4560d14002f30fb5db8314ac5a3f8c1a150e93de5317f3b830eaee4772ba"} Dec 03 11:15:58 crc kubenswrapper[4646]: I1203 11:15:58.948468 4646 generic.go:334] "Generic (PLEG): container finished" podID="e5dc2d79-d598-4b4d-80fd-4ac11612b32f" containerID="1b476e700d6473b5090594015176b8ef1b20eab1116d439a30d346cb91ea20d6" exitCode=0 Dec 03 11:15:58 crc kubenswrapper[4646]: I1203 11:15:58.948517 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5dc2d79-d598-4b4d-80fd-4ac11612b32f","Type":"ContainerDied","Data":"1b476e700d6473b5090594015176b8ef1b20eab1116d439a30d346cb91ea20d6"} Dec 03 11:15:58 crc kubenswrapper[4646]: I1203 11:15:58.948543 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5dc2d79-d598-4b4d-80fd-4ac11612b32f","Type":"ContainerDied","Data":"644b3b6d9b1afb184b487eda04cb7bda8784fe805939261c15c79e4d3c254981"} Dec 03 11:15:58 crc kubenswrapper[4646]: I1203 11:15:58.948563 4646 scope.go:117] "RemoveContainer" containerID="1b476e700d6473b5090594015176b8ef1b20eab1116d439a30d346cb91ea20d6" Dec 03 11:15:58 crc kubenswrapper[4646]: I1203 11:15:58.948624 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 11:15:58 crc kubenswrapper[4646]: I1203 11:15:58.966735 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.9667132240000003 podStartE2EDuration="2.966713224s" podCreationTimestamp="2025-12-03 11:15:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:15:58.9529789 +0000 UTC m=+1335.416035035" watchObservedRunningTime="2025-12-03 11:15:58.966713224 +0000 UTC m=+1335.429769359" Dec 03 11:15:58 crc kubenswrapper[4646]: I1203 11:15:58.973834 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-config-data\") pod \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " Dec 03 11:15:58 crc kubenswrapper[4646]: I1203 11:15:58.973978 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfmnr\" (UniqueName: \"kubernetes.io/projected/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-kube-api-access-qfmnr\") pod \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " Dec 03 11:15:58 crc kubenswrapper[4646]: I1203 11:15:58.974059 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-logs\") pod \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " Dec 03 11:15:58 crc kubenswrapper[4646]: I1203 11:15:58.974104 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-nova-metadata-tls-certs\") pod \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " Dec 03 11:15:58 crc kubenswrapper[4646]: I1203 11:15:58.974152 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-combined-ca-bundle\") pod \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\" (UID: \"e5dc2d79-d598-4b4d-80fd-4ac11612b32f\") " Dec 03 11:15:58 crc kubenswrapper[4646]: I1203 11:15:58.975596 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-logs" (OuterVolumeSpecName: "logs") pod "e5dc2d79-d598-4b4d-80fd-4ac11612b32f" (UID: "e5dc2d79-d598-4b4d-80fd-4ac11612b32f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:15:58 crc kubenswrapper[4646]: I1203 11:15:58.988620 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-kube-api-access-qfmnr" (OuterVolumeSpecName: "kube-api-access-qfmnr") pod "e5dc2d79-d598-4b4d-80fd-4ac11612b32f" (UID: "e5dc2d79-d598-4b4d-80fd-4ac11612b32f"). InnerVolumeSpecName "kube-api-access-qfmnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.016584 4646 scope.go:117] "RemoveContainer" containerID="5e09820f691f7079067c490be8272a91f0bb4eec739e409bb9604ab4d889a4b4" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.051605 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5dc2d79-d598-4b4d-80fd-4ac11612b32f" (UID: "e5dc2d79-d598-4b4d-80fd-4ac11612b32f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.063669 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "e5dc2d79-d598-4b4d-80fd-4ac11612b32f" (UID: "e5dc2d79-d598-4b4d-80fd-4ac11612b32f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.064200 4646 scope.go:117] "RemoveContainer" containerID="1b476e700d6473b5090594015176b8ef1b20eab1116d439a30d346cb91ea20d6" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.064527 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-config-data" (OuterVolumeSpecName: "config-data") pod "e5dc2d79-d598-4b4d-80fd-4ac11612b32f" (UID: "e5dc2d79-d598-4b4d-80fd-4ac11612b32f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:59 crc kubenswrapper[4646]: E1203 11:15:59.065683 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b476e700d6473b5090594015176b8ef1b20eab1116d439a30d346cb91ea20d6\": container with ID starting with 1b476e700d6473b5090594015176b8ef1b20eab1116d439a30d346cb91ea20d6 not found: ID does not exist" containerID="1b476e700d6473b5090594015176b8ef1b20eab1116d439a30d346cb91ea20d6" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.065724 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b476e700d6473b5090594015176b8ef1b20eab1116d439a30d346cb91ea20d6"} err="failed to get container status \"1b476e700d6473b5090594015176b8ef1b20eab1116d439a30d346cb91ea20d6\": rpc error: code = NotFound desc = could not find container \"1b476e700d6473b5090594015176b8ef1b20eab1116d439a30d346cb91ea20d6\": container with ID starting with 1b476e700d6473b5090594015176b8ef1b20eab1116d439a30d346cb91ea20d6 not found: ID does not exist" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.065768 4646 scope.go:117] "RemoveContainer" containerID="5e09820f691f7079067c490be8272a91f0bb4eec739e409bb9604ab4d889a4b4" Dec 03 11:15:59 crc kubenswrapper[4646]: E1203 11:15:59.066039 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e09820f691f7079067c490be8272a91f0bb4eec739e409bb9604ab4d889a4b4\": container with ID starting with 5e09820f691f7079067c490be8272a91f0bb4eec739e409bb9604ab4d889a4b4 not found: ID does not exist" containerID="5e09820f691f7079067c490be8272a91f0bb4eec739e409bb9604ab4d889a4b4" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.066082 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e09820f691f7079067c490be8272a91f0bb4eec739e409bb9604ab4d889a4b4"} err="failed to get container status \"5e09820f691f7079067c490be8272a91f0bb4eec739e409bb9604ab4d889a4b4\": rpc error: code = NotFound desc = could not find container \"5e09820f691f7079067c490be8272a91f0bb4eec739e409bb9604ab4d889a4b4\": container with ID starting with 5e09820f691f7079067c490be8272a91f0bb4eec739e409bb9604ab4d889a4b4 not found: ID does not exist" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.076203 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.076227 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfmnr\" (UniqueName: \"kubernetes.io/projected/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-kube-api-access-qfmnr\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.076239 4646 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-logs\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.076263 4646 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.076271 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5dc2d79-d598-4b4d-80fd-4ac11612b32f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.318320 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.336224 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.356402 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:59 crc kubenswrapper[4646]: E1203 11:15:59.356903 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5dc2d79-d598-4b4d-80fd-4ac11612b32f" containerName="nova-metadata-log" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.356927 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5dc2d79-d598-4b4d-80fd-4ac11612b32f" containerName="nova-metadata-log" Dec 03 11:15:59 crc kubenswrapper[4646]: E1203 11:15:59.356952 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5dc2d79-d598-4b4d-80fd-4ac11612b32f" containerName="nova-metadata-metadata" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.356962 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5dc2d79-d598-4b4d-80fd-4ac11612b32f" containerName="nova-metadata-metadata" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.357167 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5dc2d79-d598-4b4d-80fd-4ac11612b32f" containerName="nova-metadata-log" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.357197 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5dc2d79-d598-4b4d-80fd-4ac11612b32f" containerName="nova-metadata-metadata" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.358387 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.363219 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.363448 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.365219 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.484018 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.486320 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/27ad9778-57c8-4308-95e8-c20592f589c7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"27ad9778-57c8-4308-95e8-c20592f589c7\") " pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.486384 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27ad9778-57c8-4308-95e8-c20592f589c7-config-data\") pod \"nova-metadata-0\" (UID: \"27ad9778-57c8-4308-95e8-c20592f589c7\") " pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.486422 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hlfh\" (UniqueName: \"kubernetes.io/projected/27ad9778-57c8-4308-95e8-c20592f589c7-kube-api-access-7hlfh\") pod \"nova-metadata-0\" (UID: \"27ad9778-57c8-4308-95e8-c20592f589c7\") " pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.486526 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27ad9778-57c8-4308-95e8-c20592f589c7-logs\") pod \"nova-metadata-0\" (UID: \"27ad9778-57c8-4308-95e8-c20592f589c7\") " pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.486558 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27ad9778-57c8-4308-95e8-c20592f589c7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"27ad9778-57c8-4308-95e8-c20592f589c7\") " pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.587858 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7a75658-d6ae-4875-a787-0b3e49697499-combined-ca-bundle\") pod \"a7a75658-d6ae-4875-a787-0b3e49697499\" (UID: \"a7a75658-d6ae-4875-a787-0b3e49697499\") " Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.587937 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnf75\" (UniqueName: \"kubernetes.io/projected/a7a75658-d6ae-4875-a787-0b3e49697499-kube-api-access-nnf75\") pod \"a7a75658-d6ae-4875-a787-0b3e49697499\" (UID: \"a7a75658-d6ae-4875-a787-0b3e49697499\") " Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.587972 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7a75658-d6ae-4875-a787-0b3e49697499-config-data\") pod \"a7a75658-d6ae-4875-a787-0b3e49697499\" (UID: \"a7a75658-d6ae-4875-a787-0b3e49697499\") " Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.588256 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27ad9778-57c8-4308-95e8-c20592f589c7-logs\") pod \"nova-metadata-0\" (UID: \"27ad9778-57c8-4308-95e8-c20592f589c7\") " pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.588303 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27ad9778-57c8-4308-95e8-c20592f589c7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"27ad9778-57c8-4308-95e8-c20592f589c7\") " pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.588351 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/27ad9778-57c8-4308-95e8-c20592f589c7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"27ad9778-57c8-4308-95e8-c20592f589c7\") " pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.588386 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27ad9778-57c8-4308-95e8-c20592f589c7-config-data\") pod \"nova-metadata-0\" (UID: \"27ad9778-57c8-4308-95e8-c20592f589c7\") " pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.588421 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hlfh\" (UniqueName: \"kubernetes.io/projected/27ad9778-57c8-4308-95e8-c20592f589c7-kube-api-access-7hlfh\") pod \"nova-metadata-0\" (UID: \"27ad9778-57c8-4308-95e8-c20592f589c7\") " pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.588860 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27ad9778-57c8-4308-95e8-c20592f589c7-logs\") pod \"nova-metadata-0\" (UID: \"27ad9778-57c8-4308-95e8-c20592f589c7\") " pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.596852 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7a75658-d6ae-4875-a787-0b3e49697499-kube-api-access-nnf75" (OuterVolumeSpecName: "kube-api-access-nnf75") pod "a7a75658-d6ae-4875-a787-0b3e49697499" (UID: "a7a75658-d6ae-4875-a787-0b3e49697499"). InnerVolumeSpecName "kube-api-access-nnf75". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.597386 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27ad9778-57c8-4308-95e8-c20592f589c7-config-data\") pod \"nova-metadata-0\" (UID: \"27ad9778-57c8-4308-95e8-c20592f589c7\") " pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.597500 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27ad9778-57c8-4308-95e8-c20592f589c7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"27ad9778-57c8-4308-95e8-c20592f589c7\") " pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.610245 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/27ad9778-57c8-4308-95e8-c20592f589c7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"27ad9778-57c8-4308-95e8-c20592f589c7\") " pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.610641 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hlfh\" (UniqueName: \"kubernetes.io/projected/27ad9778-57c8-4308-95e8-c20592f589c7-kube-api-access-7hlfh\") pod \"nova-metadata-0\" (UID: \"27ad9778-57c8-4308-95e8-c20592f589c7\") " pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.627438 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7a75658-d6ae-4875-a787-0b3e49697499-config-data" (OuterVolumeSpecName: "config-data") pod "a7a75658-d6ae-4875-a787-0b3e49697499" (UID: "a7a75658-d6ae-4875-a787-0b3e49697499"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.645617 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7a75658-d6ae-4875-a787-0b3e49697499-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7a75658-d6ae-4875-a787-0b3e49697499" (UID: "a7a75658-d6ae-4875-a787-0b3e49697499"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.680233 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.692201 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnf75\" (UniqueName: \"kubernetes.io/projected/a7a75658-d6ae-4875-a787-0b3e49697499-kube-api-access-nnf75\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.692268 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7a75658-d6ae-4875-a787-0b3e49697499-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.692283 4646 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7a75658-d6ae-4875-a787-0b3e49697499-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.866404 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5dc2d79-d598-4b4d-80fd-4ac11612b32f" path="/var/lib/kubelet/pods/e5dc2d79-d598-4b4d-80fd-4ac11612b32f/volumes" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.969419 4646 generic.go:334] "Generic (PLEG): container finished" podID="a7a75658-d6ae-4875-a787-0b3e49697499" containerID="5265a31fb663da564a289470d89f019e099cf570cad86d4d793f454e0ede77fa" exitCode=0 Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.970245 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.970638 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a7a75658-d6ae-4875-a787-0b3e49697499","Type":"ContainerDied","Data":"5265a31fb663da564a289470d89f019e099cf570cad86d4d793f454e0ede77fa"} Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.970709 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a7a75658-d6ae-4875-a787-0b3e49697499","Type":"ContainerDied","Data":"57d8de499733cc2e0dc409603316ec1598b6fc351ac8a1e5cc2429f4b5c5a4ff"} Dec 03 11:15:59 crc kubenswrapper[4646]: I1203 11:15:59.970726 4646 scope.go:117] "RemoveContainer" containerID="5265a31fb663da564a289470d89f019e099cf570cad86d4d793f454e0ede77fa" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.006390 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.007166 4646 scope.go:117] "RemoveContainer" containerID="5265a31fb663da564a289470d89f019e099cf570cad86d4d793f454e0ede77fa" Dec 03 11:16:00 crc kubenswrapper[4646]: E1203 11:16:00.009074 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5265a31fb663da564a289470d89f019e099cf570cad86d4d793f454e0ede77fa\": container with ID starting with 5265a31fb663da564a289470d89f019e099cf570cad86d4d793f454e0ede77fa not found: ID does not exist" containerID="5265a31fb663da564a289470d89f019e099cf570cad86d4d793f454e0ede77fa" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.009134 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5265a31fb663da564a289470d89f019e099cf570cad86d4d793f454e0ede77fa"} err="failed to get container status \"5265a31fb663da564a289470d89f019e099cf570cad86d4d793f454e0ede77fa\": rpc error: code = NotFound desc = could not find container \"5265a31fb663da564a289470d89f019e099cf570cad86d4d793f454e0ede77fa\": container with ID starting with 5265a31fb663da564a289470d89f019e099cf570cad86d4d793f454e0ede77fa not found: ID does not exist" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.023957 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.033395 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 11:16:00 crc kubenswrapper[4646]: E1203 11:16:00.033932 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7a75658-d6ae-4875-a787-0b3e49697499" containerName="nova-scheduler-scheduler" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.033957 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7a75658-d6ae-4875-a787-0b3e49697499" containerName="nova-scheduler-scheduler" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.034192 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7a75658-d6ae-4875-a787-0b3e49697499" containerName="nova-scheduler-scheduler" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.034963 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.037613 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.040829 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.105053 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7bfc0c4-44a4-4f09-8a04-22fad18a1e69-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f7bfc0c4-44a4-4f09-8a04-22fad18a1e69\") " pod="openstack/nova-scheduler-0" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.105473 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5mrx\" (UniqueName: \"kubernetes.io/projected/f7bfc0c4-44a4-4f09-8a04-22fad18a1e69-kube-api-access-l5mrx\") pod \"nova-scheduler-0\" (UID: \"f7bfc0c4-44a4-4f09-8a04-22fad18a1e69\") " pod="openstack/nova-scheduler-0" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.105650 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7bfc0c4-44a4-4f09-8a04-22fad18a1e69-config-data\") pod \"nova-scheduler-0\" (UID: \"f7bfc0c4-44a4-4f09-8a04-22fad18a1e69\") " pod="openstack/nova-scheduler-0" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.119671 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.207260 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5mrx\" (UniqueName: \"kubernetes.io/projected/f7bfc0c4-44a4-4f09-8a04-22fad18a1e69-kube-api-access-l5mrx\") pod \"nova-scheduler-0\" (UID: \"f7bfc0c4-44a4-4f09-8a04-22fad18a1e69\") " pod="openstack/nova-scheduler-0" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.207699 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7bfc0c4-44a4-4f09-8a04-22fad18a1e69-config-data\") pod \"nova-scheduler-0\" (UID: \"f7bfc0c4-44a4-4f09-8a04-22fad18a1e69\") " pod="openstack/nova-scheduler-0" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.207900 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7bfc0c4-44a4-4f09-8a04-22fad18a1e69-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f7bfc0c4-44a4-4f09-8a04-22fad18a1e69\") " pod="openstack/nova-scheduler-0" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.214582 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7bfc0c4-44a4-4f09-8a04-22fad18a1e69-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f7bfc0c4-44a4-4f09-8a04-22fad18a1e69\") " pod="openstack/nova-scheduler-0" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.217834 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7bfc0c4-44a4-4f09-8a04-22fad18a1e69-config-data\") pod \"nova-scheduler-0\" (UID: \"f7bfc0c4-44a4-4f09-8a04-22fad18a1e69\") " pod="openstack/nova-scheduler-0" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.227657 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5mrx\" (UniqueName: \"kubernetes.io/projected/f7bfc0c4-44a4-4f09-8a04-22fad18a1e69-kube-api-access-l5mrx\") pod \"nova-scheduler-0\" (UID: \"f7bfc0c4-44a4-4f09-8a04-22fad18a1e69\") " pod="openstack/nova-scheduler-0" Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.352530 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 11:16:00 crc kubenswrapper[4646]: W1203 11:16:00.815537 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7bfc0c4_44a4_4f09_8a04_22fad18a1e69.slice/crio-24fd3f5605a522e8db16c28b1020a085cd071b80c75c62193fb5c9a132412b1b WatchSource:0}: Error finding container 24fd3f5605a522e8db16c28b1020a085cd071b80c75c62193fb5c9a132412b1b: Status 404 returned error can't find the container with id 24fd3f5605a522e8db16c28b1020a085cd071b80c75c62193fb5c9a132412b1b Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.816531 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.979318 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7bfc0c4-44a4-4f09-8a04-22fad18a1e69","Type":"ContainerStarted","Data":"24fd3f5605a522e8db16c28b1020a085cd071b80c75c62193fb5c9a132412b1b"} Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.982973 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"27ad9778-57c8-4308-95e8-c20592f589c7","Type":"ContainerStarted","Data":"2e37bff728348c0e4660b40710188ee96716623b16e37ce1cd334faab40ff1a2"} Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.983035 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"27ad9778-57c8-4308-95e8-c20592f589c7","Type":"ContainerStarted","Data":"460537c9b972cb25229f2eb93dce02c31c07872affa751dd67730da62d54860f"} Dec 03 11:16:00 crc kubenswrapper[4646]: I1203 11:16:00.983047 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"27ad9778-57c8-4308-95e8-c20592f589c7","Type":"ContainerStarted","Data":"1e9115bdefa8a8f6a4610fe2bc2bf4b97d215f0a562762945645bd86dbb0fab8"} Dec 03 11:16:01 crc kubenswrapper[4646]: I1203 11:16:01.001365 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.001350699 podStartE2EDuration="2.001350699s" podCreationTimestamp="2025-12-03 11:15:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:16:00.998203661 +0000 UTC m=+1337.461259796" watchObservedRunningTime="2025-12-03 11:16:01.001350699 +0000 UTC m=+1337.464406844" Dec 03 11:16:01 crc kubenswrapper[4646]: I1203 11:16:01.867810 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7a75658-d6ae-4875-a787-0b3e49697499" path="/var/lib/kubelet/pods/a7a75658-d6ae-4875-a787-0b3e49697499/volumes" Dec 03 11:16:02 crc kubenswrapper[4646]: I1203 11:16:02.038026 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7bfc0c4-44a4-4f09-8a04-22fad18a1e69","Type":"ContainerStarted","Data":"4d198d781285a9acb840c316a52921b28562653e70567bcdddb9f1a770e2f30e"} Dec 03 11:16:02 crc kubenswrapper[4646]: I1203 11:16:02.065201 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.065183034 podStartE2EDuration="3.065183034s" podCreationTimestamp="2025-12-03 11:15:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:16:02.064839315 +0000 UTC m=+1338.527895460" watchObservedRunningTime="2025-12-03 11:16:02.065183034 +0000 UTC m=+1338.528239179" Dec 03 11:16:04 crc kubenswrapper[4646]: I1203 11:16:04.681686 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 11:16:04 crc kubenswrapper[4646]: I1203 11:16:04.681979 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 11:16:05 crc kubenswrapper[4646]: I1203 11:16:05.352858 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 11:16:07 crc kubenswrapper[4646]: I1203 11:16:07.282300 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 11:16:07 crc kubenswrapper[4646]: I1203 11:16:07.282884 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 11:16:08 crc kubenswrapper[4646]: I1203 11:16:08.301580 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="99c9b16d-acd9-453f-adb9-505992d25ec2" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.186:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 11:16:08 crc kubenswrapper[4646]: I1203 11:16:08.302098 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="99c9b16d-acd9-453f-adb9-505992d25ec2" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 11:16:09 crc kubenswrapper[4646]: I1203 11:16:09.682031 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 11:16:09 crc kubenswrapper[4646]: I1203 11:16:09.682387 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 11:16:10 crc kubenswrapper[4646]: I1203 11:16:10.353974 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 11:16:10 crc kubenswrapper[4646]: I1203 11:16:10.390592 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 11:16:10 crc kubenswrapper[4646]: I1203 11:16:10.695530 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="27ad9778-57c8-4308-95e8-c20592f589c7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.187:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 11:16:10 crc kubenswrapper[4646]: I1203 11:16:10.695559 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="27ad9778-57c8-4308-95e8-c20592f589c7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.187:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 11:16:11 crc kubenswrapper[4646]: I1203 11:16:11.149805 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 11:16:13 crc kubenswrapper[4646]: I1203 11:16:13.377210 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 11:16:17 crc kubenswrapper[4646]: I1203 11:16:17.288868 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 11:16:17 crc kubenswrapper[4646]: I1203 11:16:17.289451 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 11:16:17 crc kubenswrapper[4646]: I1203 11:16:17.289905 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 11:16:17 crc kubenswrapper[4646]: I1203 11:16:17.290016 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 11:16:17 crc kubenswrapper[4646]: I1203 11:16:17.296402 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 11:16:17 crc kubenswrapper[4646]: I1203 11:16:17.296971 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 11:16:19 crc kubenswrapper[4646]: I1203 11:16:19.687576 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 11:16:19 crc kubenswrapper[4646]: I1203 11:16:19.687930 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 11:16:19 crc kubenswrapper[4646]: I1203 11:16:19.693260 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 11:16:19 crc kubenswrapper[4646]: I1203 11:16:19.693571 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 11:16:27 crc kubenswrapper[4646]: I1203 11:16:27.957839 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 11:16:28 crc kubenswrapper[4646]: I1203 11:16:28.805453 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 11:16:32 crc kubenswrapper[4646]: I1203 11:16:32.643752 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="26f0a162-8af0-438b-b48e-5bb76cb9a645" containerName="rabbitmq" containerID="cri-o://98c8221e6383dc19e35b0d441fe752daa788c1ed0ed5b4a1b5eb71276729f617" gracePeriod=604796 Dec 03 11:16:33 crc kubenswrapper[4646]: I1203 11:16:33.256849 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="a6ab009e-7121-43c2-9471-ea5145e118d6" containerName="rabbitmq" containerID="cri-o://6a510054cfcd3acdae0fec82f5d96fad1810c4ce06e56cf7b626b54fe2b008dd" gracePeriod=604796 Dec 03 11:16:36 crc kubenswrapper[4646]: I1203 11:16:36.805753 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="26f0a162-8af0-438b-b48e-5bb76cb9a645" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 03 11:16:37 crc kubenswrapper[4646]: I1203 11:16:37.302932 4646 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="a6ab009e-7121-43c2-9471-ea5145e118d6" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.247303 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.279272 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gqgx\" (UniqueName: \"kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-kube-api-access-5gqgx\") pod \"26f0a162-8af0-438b-b48e-5bb76cb9a645\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.279342 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-confd\") pod \"26f0a162-8af0-438b-b48e-5bb76cb9a645\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.279379 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-config-data\") pod \"26f0a162-8af0-438b-b48e-5bb76cb9a645\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.279424 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-plugins\") pod \"26f0a162-8af0-438b-b48e-5bb76cb9a645\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.279476 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"26f0a162-8af0-438b-b48e-5bb76cb9a645\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.279889 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "26f0a162-8af0-438b-b48e-5bb76cb9a645" (UID: "26f0a162-8af0-438b-b48e-5bb76cb9a645"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.279942 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-server-conf\") pod \"26f0a162-8af0-438b-b48e-5bb76cb9a645\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.279968 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/26f0a162-8af0-438b-b48e-5bb76cb9a645-erlang-cookie-secret\") pod \"26f0a162-8af0-438b-b48e-5bb76cb9a645\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.279994 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/26f0a162-8af0-438b-b48e-5bb76cb9a645-pod-info\") pod \"26f0a162-8af0-438b-b48e-5bb76cb9a645\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.280351 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-plugins-conf\") pod \"26f0a162-8af0-438b-b48e-5bb76cb9a645\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.280393 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-erlang-cookie\") pod \"26f0a162-8af0-438b-b48e-5bb76cb9a645\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.280449 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-tls\") pod \"26f0a162-8af0-438b-b48e-5bb76cb9a645\" (UID: \"26f0a162-8af0-438b-b48e-5bb76cb9a645\") " Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.280972 4646 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.292115 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "26f0a162-8af0-438b-b48e-5bb76cb9a645" (UID: "26f0a162-8af0-438b-b48e-5bb76cb9a645"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.292743 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "26f0a162-8af0-438b-b48e-5bb76cb9a645" (UID: "26f0a162-8af0-438b-b48e-5bb76cb9a645"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.295727 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "26f0a162-8af0-438b-b48e-5bb76cb9a645" (UID: "26f0a162-8af0-438b-b48e-5bb76cb9a645"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.299114 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-kube-api-access-5gqgx" (OuterVolumeSpecName: "kube-api-access-5gqgx") pod "26f0a162-8af0-438b-b48e-5bb76cb9a645" (UID: "26f0a162-8af0-438b-b48e-5bb76cb9a645"). InnerVolumeSpecName "kube-api-access-5gqgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.324621 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26f0a162-8af0-438b-b48e-5bb76cb9a645-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "26f0a162-8af0-438b-b48e-5bb76cb9a645" (UID: "26f0a162-8af0-438b-b48e-5bb76cb9a645"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.324716 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/26f0a162-8af0-438b-b48e-5bb76cb9a645-pod-info" (OuterVolumeSpecName: "pod-info") pod "26f0a162-8af0-438b-b48e-5bb76cb9a645" (UID: "26f0a162-8af0-438b-b48e-5bb76cb9a645"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.325068 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "26f0a162-8af0-438b-b48e-5bb76cb9a645" (UID: "26f0a162-8af0-438b-b48e-5bb76cb9a645"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.343170 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-config-data" (OuterVolumeSpecName: "config-data") pod "26f0a162-8af0-438b-b48e-5bb76cb9a645" (UID: "26f0a162-8af0-438b-b48e-5bb76cb9a645"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.371003 4646 generic.go:334] "Generic (PLEG): container finished" podID="26f0a162-8af0-438b-b48e-5bb76cb9a645" containerID="98c8221e6383dc19e35b0d441fe752daa788c1ed0ed5b4a1b5eb71276729f617" exitCode=0 Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.371054 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"26f0a162-8af0-438b-b48e-5bb76cb9a645","Type":"ContainerDied","Data":"98c8221e6383dc19e35b0d441fe752daa788c1ed0ed5b4a1b5eb71276729f617"} Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.371086 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"26f0a162-8af0-438b-b48e-5bb76cb9a645","Type":"ContainerDied","Data":"f8b457d8b1350ba31368c1d3d533fc52d276eb34641e7e28080ea22454b88056"} Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.371113 4646 scope.go:117] "RemoveContainer" containerID="98c8221e6383dc19e35b0d441fe752daa788c1ed0ed5b4a1b5eb71276729f617" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.371313 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.390737 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gqgx\" (UniqueName: \"kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-kube-api-access-5gqgx\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.390774 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.390801 4646 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.390812 4646 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/26f0a162-8af0-438b-b48e-5bb76cb9a645-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.390823 4646 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/26f0a162-8af0-438b-b48e-5bb76cb9a645-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.390845 4646 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.390855 4646 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.390865 4646 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.392026 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-server-conf" (OuterVolumeSpecName: "server-conf") pod "26f0a162-8af0-438b-b48e-5bb76cb9a645" (UID: "26f0a162-8af0-438b-b48e-5bb76cb9a645"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.433277 4646 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.456024 4646 scope.go:117] "RemoveContainer" containerID="403c513e6091509f7524e462220d0d7d652a7b5bfdbdd96c35cd55a4ea200891" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.493649 4646 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.493682 4646 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/26f0a162-8af0-438b-b48e-5bb76cb9a645-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.523536 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "26f0a162-8af0-438b-b48e-5bb76cb9a645" (UID: "26f0a162-8af0-438b-b48e-5bb76cb9a645"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.524899 4646 scope.go:117] "RemoveContainer" containerID="98c8221e6383dc19e35b0d441fe752daa788c1ed0ed5b4a1b5eb71276729f617" Dec 03 11:16:39 crc kubenswrapper[4646]: E1203 11:16:39.525432 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98c8221e6383dc19e35b0d441fe752daa788c1ed0ed5b4a1b5eb71276729f617\": container with ID starting with 98c8221e6383dc19e35b0d441fe752daa788c1ed0ed5b4a1b5eb71276729f617 not found: ID does not exist" containerID="98c8221e6383dc19e35b0d441fe752daa788c1ed0ed5b4a1b5eb71276729f617" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.525487 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98c8221e6383dc19e35b0d441fe752daa788c1ed0ed5b4a1b5eb71276729f617"} err="failed to get container status \"98c8221e6383dc19e35b0d441fe752daa788c1ed0ed5b4a1b5eb71276729f617\": rpc error: code = NotFound desc = could not find container \"98c8221e6383dc19e35b0d441fe752daa788c1ed0ed5b4a1b5eb71276729f617\": container with ID starting with 98c8221e6383dc19e35b0d441fe752daa788c1ed0ed5b4a1b5eb71276729f617 not found: ID does not exist" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.525513 4646 scope.go:117] "RemoveContainer" containerID="403c513e6091509f7524e462220d0d7d652a7b5bfdbdd96c35cd55a4ea200891" Dec 03 11:16:39 crc kubenswrapper[4646]: E1203 11:16:39.525831 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"403c513e6091509f7524e462220d0d7d652a7b5bfdbdd96c35cd55a4ea200891\": container with ID starting with 403c513e6091509f7524e462220d0d7d652a7b5bfdbdd96c35cd55a4ea200891 not found: ID does not exist" containerID="403c513e6091509f7524e462220d0d7d652a7b5bfdbdd96c35cd55a4ea200891" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.525859 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"403c513e6091509f7524e462220d0d7d652a7b5bfdbdd96c35cd55a4ea200891"} err="failed to get container status \"403c513e6091509f7524e462220d0d7d652a7b5bfdbdd96c35cd55a4ea200891\": rpc error: code = NotFound desc = could not find container \"403c513e6091509f7524e462220d0d7d652a7b5bfdbdd96c35cd55a4ea200891\": container with ID starting with 403c513e6091509f7524e462220d0d7d652a7b5bfdbdd96c35cd55a4ea200891 not found: ID does not exist" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.595283 4646 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/26f0a162-8af0-438b-b48e-5bb76cb9a645-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.734840 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.748835 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.784266 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 11:16:39 crc kubenswrapper[4646]: E1203 11:16:39.784713 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f0a162-8af0-438b-b48e-5bb76cb9a645" containerName="setup-container" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.784725 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f0a162-8af0-438b-b48e-5bb76cb9a645" containerName="setup-container" Dec 03 11:16:39 crc kubenswrapper[4646]: E1203 11:16:39.784747 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f0a162-8af0-438b-b48e-5bb76cb9a645" containerName="rabbitmq" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.784753 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f0a162-8af0-438b-b48e-5bb76cb9a645" containerName="rabbitmq" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.784905 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="26f0a162-8af0-438b-b48e-5bb76cb9a645" containerName="rabbitmq" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.790302 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.795183 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.798102 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.798477 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xdhxg" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.798954 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.799296 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.799371 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.799790 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.799997 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.854241 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:39 crc kubenswrapper[4646]: I1203 11:16:39.861049 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26f0a162-8af0-438b-b48e-5bb76cb9a645" path="/var/lib/kubelet/pods/26f0a162-8af0-438b-b48e-5bb76cb9a645/volumes" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.001860 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"a6ab009e-7121-43c2-9471-ea5145e118d6\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.002209 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-plugins\") pod \"a6ab009e-7121-43c2-9471-ea5145e118d6\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.002252 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-tls\") pod \"a6ab009e-7121-43c2-9471-ea5145e118d6\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.002283 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-server-conf\") pod \"a6ab009e-7121-43c2-9471-ea5145e118d6\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.002377 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a6ab009e-7121-43c2-9471-ea5145e118d6-erlang-cookie-secret\") pod \"a6ab009e-7121-43c2-9471-ea5145e118d6\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.002410 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-config-data\") pod \"a6ab009e-7121-43c2-9471-ea5145e118d6\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.002467 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-confd\") pod \"a6ab009e-7121-43c2-9471-ea5145e118d6\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.002502 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggj5c\" (UniqueName: \"kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-kube-api-access-ggj5c\") pod \"a6ab009e-7121-43c2-9471-ea5145e118d6\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.002532 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-plugins-conf\") pod \"a6ab009e-7121-43c2-9471-ea5145e118d6\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.002572 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-erlang-cookie\") pod \"a6ab009e-7121-43c2-9471-ea5145e118d6\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.002602 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a6ab009e-7121-43c2-9471-ea5145e118d6-pod-info\") pod \"a6ab009e-7121-43c2-9471-ea5145e118d6\" (UID: \"a6ab009e-7121-43c2-9471-ea5145e118d6\") " Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.002708 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a6ab009e-7121-43c2-9471-ea5145e118d6" (UID: "a6ab009e-7121-43c2-9471-ea5145e118d6"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.002876 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3e83310b-1692-4be3-83ba-8278d0dc31e2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.002921 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3e83310b-1692-4be3-83ba-8278d0dc31e2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.002955 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3e83310b-1692-4be3-83ba-8278d0dc31e2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.003019 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99rr8\" (UniqueName: \"kubernetes.io/projected/3e83310b-1692-4be3-83ba-8278d0dc31e2-kube-api-access-99rr8\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.003053 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3e83310b-1692-4be3-83ba-8278d0dc31e2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.003350 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a6ab009e-7121-43c2-9471-ea5145e118d6" (UID: "a6ab009e-7121-43c2-9471-ea5145e118d6"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.003495 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3e83310b-1692-4be3-83ba-8278d0dc31e2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.003576 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3e83310b-1692-4be3-83ba-8278d0dc31e2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.003624 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3e83310b-1692-4be3-83ba-8278d0dc31e2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.003709 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e83310b-1692-4be3-83ba-8278d0dc31e2-config-data\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.003798 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3e83310b-1692-4be3-83ba-8278d0dc31e2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.003861 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.004590 4646 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.004623 4646 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.011464 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a6ab009e-7121-43c2-9471-ea5145e118d6-pod-info" (OuterVolumeSpecName: "pod-info") pod "a6ab009e-7121-43c2-9471-ea5145e118d6" (UID: "a6ab009e-7121-43c2-9471-ea5145e118d6"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.012612 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a6ab009e-7121-43c2-9471-ea5145e118d6" (UID: "a6ab009e-7121-43c2-9471-ea5145e118d6"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.013048 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a6ab009e-7121-43c2-9471-ea5145e118d6" (UID: "a6ab009e-7121-43c2-9471-ea5145e118d6"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.015534 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "a6ab009e-7121-43c2-9471-ea5145e118d6" (UID: "a6ab009e-7121-43c2-9471-ea5145e118d6"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.021107 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-kube-api-access-ggj5c" (OuterVolumeSpecName: "kube-api-access-ggj5c") pod "a6ab009e-7121-43c2-9471-ea5145e118d6" (UID: "a6ab009e-7121-43c2-9471-ea5145e118d6"). InnerVolumeSpecName "kube-api-access-ggj5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.024616 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6ab009e-7121-43c2-9471-ea5145e118d6-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a6ab009e-7121-43c2-9471-ea5145e118d6" (UID: "a6ab009e-7121-43c2-9471-ea5145e118d6"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.039901 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-config-data" (OuterVolumeSpecName: "config-data") pod "a6ab009e-7121-43c2-9471-ea5145e118d6" (UID: "a6ab009e-7121-43c2-9471-ea5145e118d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.075972 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-server-conf" (OuterVolumeSpecName: "server-conf") pod "a6ab009e-7121-43c2-9471-ea5145e118d6" (UID: "a6ab009e-7121-43c2-9471-ea5145e118d6"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106158 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3e83310b-1692-4be3-83ba-8278d0dc31e2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106227 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3e83310b-1692-4be3-83ba-8278d0dc31e2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106279 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e83310b-1692-4be3-83ba-8278d0dc31e2-config-data\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106313 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3e83310b-1692-4be3-83ba-8278d0dc31e2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106361 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106404 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3e83310b-1692-4be3-83ba-8278d0dc31e2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106424 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3e83310b-1692-4be3-83ba-8278d0dc31e2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106441 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3e83310b-1692-4be3-83ba-8278d0dc31e2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106479 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99rr8\" (UniqueName: \"kubernetes.io/projected/3e83310b-1692-4be3-83ba-8278d0dc31e2-kube-api-access-99rr8\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106498 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3e83310b-1692-4be3-83ba-8278d0dc31e2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106514 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3e83310b-1692-4be3-83ba-8278d0dc31e2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106587 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggj5c\" (UniqueName: \"kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-kube-api-access-ggj5c\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106598 4646 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106606 4646 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a6ab009e-7121-43c2-9471-ea5145e118d6-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106657 4646 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106666 4646 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106674 4646 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106682 4646 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a6ab009e-7121-43c2-9471-ea5145e118d6-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.106691 4646 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6ab009e-7121-43c2-9471-ea5145e118d6-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.107682 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e83310b-1692-4be3-83ba-8278d0dc31e2-config-data\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.108003 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3e83310b-1692-4be3-83ba-8278d0dc31e2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.108292 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3e83310b-1692-4be3-83ba-8278d0dc31e2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.108960 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3e83310b-1692-4be3-83ba-8278d0dc31e2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.109163 4646 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.109186 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3e83310b-1692-4be3-83ba-8278d0dc31e2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.121571 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3e83310b-1692-4be3-83ba-8278d0dc31e2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.122120 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3e83310b-1692-4be3-83ba-8278d0dc31e2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.122564 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3e83310b-1692-4be3-83ba-8278d0dc31e2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.127561 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3e83310b-1692-4be3-83ba-8278d0dc31e2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.129738 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99rr8\" (UniqueName: \"kubernetes.io/projected/3e83310b-1692-4be3-83ba-8278d0dc31e2-kube-api-access-99rr8\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.148875 4646 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.170423 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"3e83310b-1692-4be3-83ba-8278d0dc31e2\") " pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.182947 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a6ab009e-7121-43c2-9471-ea5145e118d6" (UID: "a6ab009e-7121-43c2-9471-ea5145e118d6"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.209240 4646 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.209294 4646 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a6ab009e-7121-43c2-9471-ea5145e118d6-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.432628 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.439393 4646 generic.go:334] "Generic (PLEG): container finished" podID="a6ab009e-7121-43c2-9471-ea5145e118d6" containerID="6a510054cfcd3acdae0fec82f5d96fad1810c4ce06e56cf7b626b54fe2b008dd" exitCode=0 Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.439455 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a6ab009e-7121-43c2-9471-ea5145e118d6","Type":"ContainerDied","Data":"6a510054cfcd3acdae0fec82f5d96fad1810c4ce06e56cf7b626b54fe2b008dd"} Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.439490 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a6ab009e-7121-43c2-9471-ea5145e118d6","Type":"ContainerDied","Data":"fa3f6e5219fc49110caa3898635ec4817eb17ab14902cde64a0ec5ff367058ef"} Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.439513 4646 scope.go:117] "RemoveContainer" containerID="6a510054cfcd3acdae0fec82f5d96fad1810c4ce06e56cf7b626b54fe2b008dd" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.439538 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.497236 4646 scope.go:117] "RemoveContainer" containerID="c5a483953bdf195796f57d4f668b4cea83e28595a29fc5705a6029cc8dbf8bf3" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.500802 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.528397 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.550783 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 11:16:40 crc kubenswrapper[4646]: E1203 11:16:40.551255 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6ab009e-7121-43c2-9471-ea5145e118d6" containerName="setup-container" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.551275 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6ab009e-7121-43c2-9471-ea5145e118d6" containerName="setup-container" Dec 03 11:16:40 crc kubenswrapper[4646]: E1203 11:16:40.551287 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6ab009e-7121-43c2-9471-ea5145e118d6" containerName="rabbitmq" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.551294 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6ab009e-7121-43c2-9471-ea5145e118d6" containerName="rabbitmq" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.551628 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6ab009e-7121-43c2-9471-ea5145e118d6" containerName="rabbitmq" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.553641 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.556930 4646 scope.go:117] "RemoveContainer" containerID="6a510054cfcd3acdae0fec82f5d96fad1810c4ce06e56cf7b626b54fe2b008dd" Dec 03 11:16:40 crc kubenswrapper[4646]: E1203 11:16:40.562854 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a510054cfcd3acdae0fec82f5d96fad1810c4ce06e56cf7b626b54fe2b008dd\": container with ID starting with 6a510054cfcd3acdae0fec82f5d96fad1810c4ce06e56cf7b626b54fe2b008dd not found: ID does not exist" containerID="6a510054cfcd3acdae0fec82f5d96fad1810c4ce06e56cf7b626b54fe2b008dd" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.562885 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a510054cfcd3acdae0fec82f5d96fad1810c4ce06e56cf7b626b54fe2b008dd"} err="failed to get container status \"6a510054cfcd3acdae0fec82f5d96fad1810c4ce06e56cf7b626b54fe2b008dd\": rpc error: code = NotFound desc = could not find container \"6a510054cfcd3acdae0fec82f5d96fad1810c4ce06e56cf7b626b54fe2b008dd\": container with ID starting with 6a510054cfcd3acdae0fec82f5d96fad1810c4ce06e56cf7b626b54fe2b008dd not found: ID does not exist" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.562910 4646 scope.go:117] "RemoveContainer" containerID="c5a483953bdf195796f57d4f668b4cea83e28595a29fc5705a6029cc8dbf8bf3" Dec 03 11:16:40 crc kubenswrapper[4646]: E1203 11:16:40.565882 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5a483953bdf195796f57d4f668b4cea83e28595a29fc5705a6029cc8dbf8bf3\": container with ID starting with c5a483953bdf195796f57d4f668b4cea83e28595a29fc5705a6029cc8dbf8bf3 not found: ID does not exist" containerID="c5a483953bdf195796f57d4f668b4cea83e28595a29fc5705a6029cc8dbf8bf3" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.565912 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a483953bdf195796f57d4f668b4cea83e28595a29fc5705a6029cc8dbf8bf3"} err="failed to get container status \"c5a483953bdf195796f57d4f668b4cea83e28595a29fc5705a6029cc8dbf8bf3\": rpc error: code = NotFound desc = could not find container \"c5a483953bdf195796f57d4f668b4cea83e28595a29fc5705a6029cc8dbf8bf3\": container with ID starting with c5a483953bdf195796f57d4f668b4cea83e28595a29fc5705a6029cc8dbf8bf3 not found: ID does not exist" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.568691 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.568991 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.569175 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-c4pm6" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.569349 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.569497 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.569621 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.573360 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.576260 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.736107 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.736493 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/16b4b1d1-a293-4042-b52b-dc8a10e07536-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.736532 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/16b4b1d1-a293-4042-b52b-dc8a10e07536-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.736563 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/16b4b1d1-a293-4042-b52b-dc8a10e07536-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.736584 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/16b4b1d1-a293-4042-b52b-dc8a10e07536-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.736606 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/16b4b1d1-a293-4042-b52b-dc8a10e07536-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.736646 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16b4b1d1-a293-4042-b52b-dc8a10e07536-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.736673 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xfks\" (UniqueName: \"kubernetes.io/projected/16b4b1d1-a293-4042-b52b-dc8a10e07536-kube-api-access-4xfks\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.736773 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/16b4b1d1-a293-4042-b52b-dc8a10e07536-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.736802 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/16b4b1d1-a293-4042-b52b-dc8a10e07536-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.736832 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/16b4b1d1-a293-4042-b52b-dc8a10e07536-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.838914 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/16b4b1d1-a293-4042-b52b-dc8a10e07536-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.838963 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/16b4b1d1-a293-4042-b52b-dc8a10e07536-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.839017 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.839063 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/16b4b1d1-a293-4042-b52b-dc8a10e07536-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.839090 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/16b4b1d1-a293-4042-b52b-dc8a10e07536-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.839111 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/16b4b1d1-a293-4042-b52b-dc8a10e07536-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.839133 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/16b4b1d1-a293-4042-b52b-dc8a10e07536-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.839156 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/16b4b1d1-a293-4042-b52b-dc8a10e07536-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.839721 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/16b4b1d1-a293-4042-b52b-dc8a10e07536-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.839783 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16b4b1d1-a293-4042-b52b-dc8a10e07536-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.839827 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xfks\" (UniqueName: \"kubernetes.io/projected/16b4b1d1-a293-4042-b52b-dc8a10e07536-kube-api-access-4xfks\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.840004 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/16b4b1d1-a293-4042-b52b-dc8a10e07536-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.840493 4646 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.840567 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/16b4b1d1-a293-4042-b52b-dc8a10e07536-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.840728 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/16b4b1d1-a293-4042-b52b-dc8a10e07536-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.840890 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/16b4b1d1-a293-4042-b52b-dc8a10e07536-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.841412 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16b4b1d1-a293-4042-b52b-dc8a10e07536-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.845952 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/16b4b1d1-a293-4042-b52b-dc8a10e07536-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.845996 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/16b4b1d1-a293-4042-b52b-dc8a10e07536-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.849457 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/16b4b1d1-a293-4042-b52b-dc8a10e07536-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.864800 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/16b4b1d1-a293-4042-b52b-dc8a10e07536-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.874047 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xfks\" (UniqueName: \"kubernetes.io/projected/16b4b1d1-a293-4042-b52b-dc8a10e07536-kube-api-access-4xfks\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.881663 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"16b4b1d1-a293-4042-b52b-dc8a10e07536\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:40 crc kubenswrapper[4646]: I1203 11:16:40.989634 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 11:16:41 crc kubenswrapper[4646]: I1203 11:16:41.179750 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:16:41 crc kubenswrapper[4646]: I1203 11:16:41.450598 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3e83310b-1692-4be3-83ba-8278d0dc31e2","Type":"ContainerStarted","Data":"f5f018e4e978d2e8315a2c50dec86e8e17627e8a80ec2a8dad67f78ce423968f"} Dec 03 11:16:41 crc kubenswrapper[4646]: I1203 11:16:41.653941 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 11:16:41 crc kubenswrapper[4646]: W1203 11:16:41.661328 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16b4b1d1_a293_4042_b52b_dc8a10e07536.slice/crio-ae41bf23b5446e3ab91b6ed3315a14a8e41e26b931659072ac710915421cbc3b WatchSource:0}: Error finding container ae41bf23b5446e3ab91b6ed3315a14a8e41e26b931659072ac710915421cbc3b: Status 404 returned error can't find the container with id ae41bf23b5446e3ab91b6ed3315a14a8e41e26b931659072ac710915421cbc3b Dec 03 11:16:41 crc kubenswrapper[4646]: I1203 11:16:41.859007 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6ab009e-7121-43c2-9471-ea5145e118d6" path="/var/lib/kubelet/pods/a6ab009e-7121-43c2-9471-ea5145e118d6/volumes" Dec 03 11:16:42 crc kubenswrapper[4646]: I1203 11:16:42.462974 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"16b4b1d1-a293-4042-b52b-dc8a10e07536","Type":"ContainerStarted","Data":"ae41bf23b5446e3ab91b6ed3315a14a8e41e26b931659072ac710915421cbc3b"} Dec 03 11:16:43 crc kubenswrapper[4646]: I1203 11:16:43.473094 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"16b4b1d1-a293-4042-b52b-dc8a10e07536","Type":"ContainerStarted","Data":"b3e56595afdbf4fa498f491dd5f6def1c362abf5289c49fe76865479c0af3207"} Dec 03 11:16:43 crc kubenswrapper[4646]: I1203 11:16:43.474780 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3e83310b-1692-4be3-83ba-8278d0dc31e2","Type":"ContainerStarted","Data":"0369de7081e874e4446d3b6f23411d7d9731c98ee08709f28c98b738b163c283"} Dec 03 11:16:43 crc kubenswrapper[4646]: I1203 11:16:43.760132 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-vwwqn"] Dec 03 11:16:43 crc kubenswrapper[4646]: I1203 11:16:43.761993 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:43 crc kubenswrapper[4646]: I1203 11:16:43.770058 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 03 11:16:43 crc kubenswrapper[4646]: I1203 11:16:43.778427 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-vwwqn"] Dec 03 11:16:43 crc kubenswrapper[4646]: I1203 11:16:43.898431 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-ovsdbserver-nb\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:43 crc kubenswrapper[4646]: I1203 11:16:43.898720 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-ovsdbserver-sb\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:43 crc kubenswrapper[4646]: I1203 11:16:43.898969 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6pbf\" (UniqueName: \"kubernetes.io/projected/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-kube-api-access-c6pbf\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:43 crc kubenswrapper[4646]: I1203 11:16:43.899093 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-config\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:43 crc kubenswrapper[4646]: I1203 11:16:43.899196 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-openstack-edpm-ipam\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:43 crc kubenswrapper[4646]: I1203 11:16:43.899325 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-dns-svc\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:44 crc kubenswrapper[4646]: I1203 11:16:44.001295 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6pbf\" (UniqueName: \"kubernetes.io/projected/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-kube-api-access-c6pbf\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:44 crc kubenswrapper[4646]: I1203 11:16:44.001363 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-config\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:44 crc kubenswrapper[4646]: I1203 11:16:44.001393 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-openstack-edpm-ipam\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:44 crc kubenswrapper[4646]: I1203 11:16:44.001414 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-dns-svc\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:44 crc kubenswrapper[4646]: I1203 11:16:44.001477 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-ovsdbserver-nb\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:44 crc kubenswrapper[4646]: I1203 11:16:44.001513 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-ovsdbserver-sb\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:44 crc kubenswrapper[4646]: I1203 11:16:44.002484 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-ovsdbserver-sb\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:44 crc kubenswrapper[4646]: I1203 11:16:44.002689 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-config\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:44 crc kubenswrapper[4646]: I1203 11:16:44.003229 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-dns-svc\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:44 crc kubenswrapper[4646]: I1203 11:16:44.003282 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-ovsdbserver-nb\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:44 crc kubenswrapper[4646]: I1203 11:16:44.003658 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 03 11:16:44 crc kubenswrapper[4646]: I1203 11:16:44.013603 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-openstack-edpm-ipam\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:44 crc kubenswrapper[4646]: I1203 11:16:44.026273 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6pbf\" (UniqueName: \"kubernetes.io/projected/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-kube-api-access-c6pbf\") pod \"dnsmasq-dns-578b8d767c-vwwqn\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:44 crc kubenswrapper[4646]: I1203 11:16:44.081314 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:44 crc kubenswrapper[4646]: I1203 11:16:44.549382 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-vwwqn"] Dec 03 11:16:44 crc kubenswrapper[4646]: W1203 11:16:44.559677 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbde97ea4_7d4e_4632_9007_6217dc8ab7ce.slice/crio-0474b888ac99b1ab835ab86de497f69d44eba907897bf5b79ff6af9603ad32ef WatchSource:0}: Error finding container 0474b888ac99b1ab835ab86de497f69d44eba907897bf5b79ff6af9603ad32ef: Status 404 returned error can't find the container with id 0474b888ac99b1ab835ab86de497f69d44eba907897bf5b79ff6af9603ad32ef Dec 03 11:16:45 crc kubenswrapper[4646]: I1203 11:16:45.491628 4646 generic.go:334] "Generic (PLEG): container finished" podID="bde97ea4-7d4e-4632-9007-6217dc8ab7ce" containerID="cea810231f6a0decd80b4a95a8ffee279df49d789a62fe5362e151be83e261e3" exitCode=0 Dec 03 11:16:45 crc kubenswrapper[4646]: I1203 11:16:45.492120 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" event={"ID":"bde97ea4-7d4e-4632-9007-6217dc8ab7ce","Type":"ContainerDied","Data":"cea810231f6a0decd80b4a95a8ffee279df49d789a62fe5362e151be83e261e3"} Dec 03 11:16:45 crc kubenswrapper[4646]: I1203 11:16:45.492150 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" event={"ID":"bde97ea4-7d4e-4632-9007-6217dc8ab7ce","Type":"ContainerStarted","Data":"0474b888ac99b1ab835ab86de497f69d44eba907897bf5b79ff6af9603ad32ef"} Dec 03 11:16:46 crc kubenswrapper[4646]: I1203 11:16:46.504533 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" event={"ID":"bde97ea4-7d4e-4632-9007-6217dc8ab7ce","Type":"ContainerStarted","Data":"28205b37d3d2366017350c24ab4bd705d4397115ff92b5f6b65f74876a25cd7f"} Dec 03 11:16:46 crc kubenswrapper[4646]: I1203 11:16:46.504875 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:46 crc kubenswrapper[4646]: I1203 11:16:46.538473 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" podStartSLOduration=3.538445726 podStartE2EDuration="3.538445726s" podCreationTimestamp="2025-12-03 11:16:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:16:46.519507586 +0000 UTC m=+1382.982563741" watchObservedRunningTime="2025-12-03 11:16:46.538445726 +0000 UTC m=+1383.001501881" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.084290 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.163093 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-5cknk"] Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.163323 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" podUID="14a94107-452f-4dc7-a96f-4cb1e0d455d4" containerName="dnsmasq-dns" containerID="cri-o://b3cd3b39c3c6c10f4ad806e0c9e17b7abbbd2895c57390aa3abfa02e54e78108" gracePeriod=10 Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.510760 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69fd9b48bc-x88c4"] Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.512580 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.533761 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69fd9b48bc-x88c4"] Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.597819 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-dns-svc\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.597859 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-openstack-edpm-ipam\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.597907 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-ovsdbserver-nb\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.597934 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-config\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.597953 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6rlj\" (UniqueName: \"kubernetes.io/projected/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-kube-api-access-m6rlj\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.598020 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-ovsdbserver-sb\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.608660 4646 generic.go:334] "Generic (PLEG): container finished" podID="14a94107-452f-4dc7-a96f-4cb1e0d455d4" containerID="b3cd3b39c3c6c10f4ad806e0c9e17b7abbbd2895c57390aa3abfa02e54e78108" exitCode=0 Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.608702 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" event={"ID":"14a94107-452f-4dc7-a96f-4cb1e0d455d4","Type":"ContainerDied","Data":"b3cd3b39c3c6c10f4ad806e0c9e17b7abbbd2895c57390aa3abfa02e54e78108"} Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.700060 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-ovsdbserver-nb\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.700112 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-config\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.700136 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6rlj\" (UniqueName: \"kubernetes.io/projected/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-kube-api-access-m6rlj\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.700238 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-ovsdbserver-sb\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.700312 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-dns-svc\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.701636 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-openstack-edpm-ipam\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.701674 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-dns-svc\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.702606 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-ovsdbserver-sb\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.702630 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-openstack-edpm-ipam\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.703300 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-ovsdbserver-nb\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.703894 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-config\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.724365 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6rlj\" (UniqueName: \"kubernetes.io/projected/b4c02b37-8c88-41d6-866f-e31ef5cdfab3-kube-api-access-m6rlj\") pod \"dnsmasq-dns-69fd9b48bc-x88c4\" (UID: \"b4c02b37-8c88-41d6-866f-e31ef5cdfab3\") " pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.832072 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:54 crc kubenswrapper[4646]: I1203 11:16:54.892794 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.008606 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-ovsdbserver-nb\") pod \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.008999 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-ovsdbserver-sb\") pod \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.009040 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-dns-svc\") pod \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.009151 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92xcj\" (UniqueName: \"kubernetes.io/projected/14a94107-452f-4dc7-a96f-4cb1e0d455d4-kube-api-access-92xcj\") pod \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.009276 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-config\") pod \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\" (UID: \"14a94107-452f-4dc7-a96f-4cb1e0d455d4\") " Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.035088 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14a94107-452f-4dc7-a96f-4cb1e0d455d4-kube-api-access-92xcj" (OuterVolumeSpecName: "kube-api-access-92xcj") pod "14a94107-452f-4dc7-a96f-4cb1e0d455d4" (UID: "14a94107-452f-4dc7-a96f-4cb1e0d455d4"). InnerVolumeSpecName "kube-api-access-92xcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.068461 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "14a94107-452f-4dc7-a96f-4cb1e0d455d4" (UID: "14a94107-452f-4dc7-a96f-4cb1e0d455d4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.070819 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-config" (OuterVolumeSpecName: "config") pod "14a94107-452f-4dc7-a96f-4cb1e0d455d4" (UID: "14a94107-452f-4dc7-a96f-4cb1e0d455d4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.079479 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "14a94107-452f-4dc7-a96f-4cb1e0d455d4" (UID: "14a94107-452f-4dc7-a96f-4cb1e0d455d4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.085902 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "14a94107-452f-4dc7-a96f-4cb1e0d455d4" (UID: "14a94107-452f-4dc7-a96f-4cb1e0d455d4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.111148 4646 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.111174 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92xcj\" (UniqueName: \"kubernetes.io/projected/14a94107-452f-4dc7-a96f-4cb1e0d455d4-kube-api-access-92xcj\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.111190 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.111201 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.111211 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a94107-452f-4dc7-a96f-4cb1e0d455d4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.326204 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69fd9b48bc-x88c4"] Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.624862 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" event={"ID":"14a94107-452f-4dc7-a96f-4cb1e0d455d4","Type":"ContainerDied","Data":"1130e383f5ebc2ab19de2432161f19994aeee5ed6043eb7fdacf3457d6c8dc00"} Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.625180 4646 scope.go:117] "RemoveContainer" containerID="b3cd3b39c3c6c10f4ad806e0c9e17b7abbbd2895c57390aa3abfa02e54e78108" Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.624875 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-5cknk" Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.627410 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" event={"ID":"b4c02b37-8c88-41d6-866f-e31ef5cdfab3","Type":"ContainerStarted","Data":"cc7b7c70756b933b9b3660a4299a6e93c769859a02a91211eb269039c0bd9752"} Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.652751 4646 scope.go:117] "RemoveContainer" containerID="215bb2e16524a7f844284d45b3e832b96911f027ca6b06a2230d5be443e886e8" Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.659139 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-5cknk"] Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.669535 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-5cknk"] Dec 03 11:16:55 crc kubenswrapper[4646]: I1203 11:16:55.859135 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14a94107-452f-4dc7-a96f-4cb1e0d455d4" path="/var/lib/kubelet/pods/14a94107-452f-4dc7-a96f-4cb1e0d455d4/volumes" Dec 03 11:16:56 crc kubenswrapper[4646]: I1203 11:16:56.641546 4646 generic.go:334] "Generic (PLEG): container finished" podID="b4c02b37-8c88-41d6-866f-e31ef5cdfab3" containerID="98e0d667e9f2541c8b5973408734c9c0124b93130a4795165725414092319e27" exitCode=0 Dec 03 11:16:56 crc kubenswrapper[4646]: I1203 11:16:56.641595 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" event={"ID":"b4c02b37-8c88-41d6-866f-e31ef5cdfab3","Type":"ContainerDied","Data":"98e0d667e9f2541c8b5973408734c9c0124b93130a4795165725414092319e27"} Dec 03 11:16:57 crc kubenswrapper[4646]: I1203 11:16:57.654482 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" event={"ID":"b4c02b37-8c88-41d6-866f-e31ef5cdfab3","Type":"ContainerStarted","Data":"2eba86ac894f4ffcaa9a8603b5c4b532af0e713da7dccb935a0486e2b2462701"} Dec 03 11:16:57 crc kubenswrapper[4646]: I1203 11:16:57.657462 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:16:57 crc kubenswrapper[4646]: I1203 11:16:57.672299 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" podStartSLOduration=3.6722799569999998 podStartE2EDuration="3.672279957s" podCreationTimestamp="2025-12-03 11:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:16:57.67129136 +0000 UTC m=+1394.134347495" watchObservedRunningTime="2025-12-03 11:16:57.672279957 +0000 UTC m=+1394.135336092" Dec 03 11:17:04 crc kubenswrapper[4646]: I1203 11:17:04.833574 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69fd9b48bc-x88c4" Dec 03 11:17:04 crc kubenswrapper[4646]: I1203 11:17:04.916906 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-vwwqn"] Dec 03 11:17:04 crc kubenswrapper[4646]: I1203 11:17:04.917168 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" podUID="bde97ea4-7d4e-4632-9007-6217dc8ab7ce" containerName="dnsmasq-dns" containerID="cri-o://28205b37d3d2366017350c24ab4bd705d4397115ff92b5f6b65f74876a25cd7f" gracePeriod=10 Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.397800 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.497993 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-ovsdbserver-nb\") pod \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.498197 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-dns-svc\") pod \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.498226 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-config\") pod \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.498284 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-ovsdbserver-sb\") pod \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.498321 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6pbf\" (UniqueName: \"kubernetes.io/projected/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-kube-api-access-c6pbf\") pod \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.498385 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-openstack-edpm-ipam\") pod \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\" (UID: \"bde97ea4-7d4e-4632-9007-6217dc8ab7ce\") " Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.508019 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-kube-api-access-c6pbf" (OuterVolumeSpecName: "kube-api-access-c6pbf") pod "bde97ea4-7d4e-4632-9007-6217dc8ab7ce" (UID: "bde97ea4-7d4e-4632-9007-6217dc8ab7ce"). InnerVolumeSpecName "kube-api-access-c6pbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.570430 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bde97ea4-7d4e-4632-9007-6217dc8ab7ce" (UID: "bde97ea4-7d4e-4632-9007-6217dc8ab7ce"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.570499 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bde97ea4-7d4e-4632-9007-6217dc8ab7ce" (UID: "bde97ea4-7d4e-4632-9007-6217dc8ab7ce"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.577061 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-config" (OuterVolumeSpecName: "config") pod "bde97ea4-7d4e-4632-9007-6217dc8ab7ce" (UID: "bde97ea4-7d4e-4632-9007-6217dc8ab7ce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.585643 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "bde97ea4-7d4e-4632-9007-6217dc8ab7ce" (UID: "bde97ea4-7d4e-4632-9007-6217dc8ab7ce"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.590039 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bde97ea4-7d4e-4632-9007-6217dc8ab7ce" (UID: "bde97ea4-7d4e-4632-9007-6217dc8ab7ce"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.600879 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.600908 4646 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.600918 4646 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-config\") on node \"crc\" DevicePath \"\"" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.600926 4646 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.600935 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6pbf\" (UniqueName: \"kubernetes.io/projected/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-kube-api-access-c6pbf\") on node \"crc\" DevicePath \"\"" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.600988 4646 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/bde97ea4-7d4e-4632-9007-6217dc8ab7ce-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.739430 4646 generic.go:334] "Generic (PLEG): container finished" podID="bde97ea4-7d4e-4632-9007-6217dc8ab7ce" containerID="28205b37d3d2366017350c24ab4bd705d4397115ff92b5f6b65f74876a25cd7f" exitCode=0 Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.739478 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" event={"ID":"bde97ea4-7d4e-4632-9007-6217dc8ab7ce","Type":"ContainerDied","Data":"28205b37d3d2366017350c24ab4bd705d4397115ff92b5f6b65f74876a25cd7f"} Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.739508 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" event={"ID":"bde97ea4-7d4e-4632-9007-6217dc8ab7ce","Type":"ContainerDied","Data":"0474b888ac99b1ab835ab86de497f69d44eba907897bf5b79ff6af9603ad32ef"} Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.739530 4646 scope.go:117] "RemoveContainer" containerID="28205b37d3d2366017350c24ab4bd705d4397115ff92b5f6b65f74876a25cd7f" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.739681 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-vwwqn" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.777524 4646 scope.go:117] "RemoveContainer" containerID="cea810231f6a0decd80b4a95a8ffee279df49d789a62fe5362e151be83e261e3" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.783676 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-vwwqn"] Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.792806 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-vwwqn"] Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.807566 4646 scope.go:117] "RemoveContainer" containerID="28205b37d3d2366017350c24ab4bd705d4397115ff92b5f6b65f74876a25cd7f" Dec 03 11:17:05 crc kubenswrapper[4646]: E1203 11:17:05.808153 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28205b37d3d2366017350c24ab4bd705d4397115ff92b5f6b65f74876a25cd7f\": container with ID starting with 28205b37d3d2366017350c24ab4bd705d4397115ff92b5f6b65f74876a25cd7f not found: ID does not exist" containerID="28205b37d3d2366017350c24ab4bd705d4397115ff92b5f6b65f74876a25cd7f" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.808197 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28205b37d3d2366017350c24ab4bd705d4397115ff92b5f6b65f74876a25cd7f"} err="failed to get container status \"28205b37d3d2366017350c24ab4bd705d4397115ff92b5f6b65f74876a25cd7f\": rpc error: code = NotFound desc = could not find container \"28205b37d3d2366017350c24ab4bd705d4397115ff92b5f6b65f74876a25cd7f\": container with ID starting with 28205b37d3d2366017350c24ab4bd705d4397115ff92b5f6b65f74876a25cd7f not found: ID does not exist" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.808224 4646 scope.go:117] "RemoveContainer" containerID="cea810231f6a0decd80b4a95a8ffee279df49d789a62fe5362e151be83e261e3" Dec 03 11:17:05 crc kubenswrapper[4646]: E1203 11:17:05.808563 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cea810231f6a0decd80b4a95a8ffee279df49d789a62fe5362e151be83e261e3\": container with ID starting with cea810231f6a0decd80b4a95a8ffee279df49d789a62fe5362e151be83e261e3 not found: ID does not exist" containerID="cea810231f6a0decd80b4a95a8ffee279df49d789a62fe5362e151be83e261e3" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.808590 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cea810231f6a0decd80b4a95a8ffee279df49d789a62fe5362e151be83e261e3"} err="failed to get container status \"cea810231f6a0decd80b4a95a8ffee279df49d789a62fe5362e151be83e261e3\": rpc error: code = NotFound desc = could not find container \"cea810231f6a0decd80b4a95a8ffee279df49d789a62fe5362e151be83e261e3\": container with ID starting with cea810231f6a0decd80b4a95a8ffee279df49d789a62fe5362e151be83e261e3 not found: ID does not exist" Dec 03 11:17:05 crc kubenswrapper[4646]: I1203 11:17:05.873445 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bde97ea4-7d4e-4632-9007-6217dc8ab7ce" path="/var/lib/kubelet/pods/bde97ea4-7d4e-4632-9007-6217dc8ab7ce/volumes" Dec 03 11:17:14 crc kubenswrapper[4646]: I1203 11:17:14.820123 4646 generic.go:334] "Generic (PLEG): container finished" podID="3e83310b-1692-4be3-83ba-8278d0dc31e2" containerID="0369de7081e874e4446d3b6f23411d7d9731c98ee08709f28c98b738b163c283" exitCode=0 Dec 03 11:17:14 crc kubenswrapper[4646]: I1203 11:17:14.820205 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3e83310b-1692-4be3-83ba-8278d0dc31e2","Type":"ContainerDied","Data":"0369de7081e874e4446d3b6f23411d7d9731c98ee08709f28c98b738b163c283"} Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.043750 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w"] Dec 03 11:17:15 crc kubenswrapper[4646]: E1203 11:17:15.044496 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bde97ea4-7d4e-4632-9007-6217dc8ab7ce" containerName="dnsmasq-dns" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.044513 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="bde97ea4-7d4e-4632-9007-6217dc8ab7ce" containerName="dnsmasq-dns" Dec 03 11:17:15 crc kubenswrapper[4646]: E1203 11:17:15.044523 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a94107-452f-4dc7-a96f-4cb1e0d455d4" containerName="dnsmasq-dns" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.044529 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a94107-452f-4dc7-a96f-4cb1e0d455d4" containerName="dnsmasq-dns" Dec 03 11:17:15 crc kubenswrapper[4646]: E1203 11:17:15.044560 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a94107-452f-4dc7-a96f-4cb1e0d455d4" containerName="init" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.044568 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a94107-452f-4dc7-a96f-4cb1e0d455d4" containerName="init" Dec 03 11:17:15 crc kubenswrapper[4646]: E1203 11:17:15.044583 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bde97ea4-7d4e-4632-9007-6217dc8ab7ce" containerName="init" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.044589 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="bde97ea4-7d4e-4632-9007-6217dc8ab7ce" containerName="init" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.044757 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="14a94107-452f-4dc7-a96f-4cb1e0d455d4" containerName="dnsmasq-dns" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.044780 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="bde97ea4-7d4e-4632-9007-6217dc8ab7ce" containerName="dnsmasq-dns" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.045513 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.048048 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.048329 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.048466 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bfp25" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.051741 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.064161 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w"] Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.141626 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnknp\" (UniqueName: \"kubernetes.io/projected/5afc0cac-1a45-4f71-94e6-241231108859-kube-api-access-rnknp\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.141695 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.141814 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.141884 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.243684 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.243757 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.244382 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnknp\" (UniqueName: \"kubernetes.io/projected/5afc0cac-1a45-4f71-94e6-241231108859-kube-api-access-rnknp\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.244444 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.247808 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.250914 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.256050 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.267209 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnknp\" (UniqueName: \"kubernetes.io/projected/5afc0cac-1a45-4f71-94e6-241231108859-kube-api-access-rnknp\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.381696 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.835980 4646 generic.go:334] "Generic (PLEG): container finished" podID="16b4b1d1-a293-4042-b52b-dc8a10e07536" containerID="b3e56595afdbf4fa498f491dd5f6def1c362abf5289c49fe76865479c0af3207" exitCode=0 Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.836293 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"16b4b1d1-a293-4042-b52b-dc8a10e07536","Type":"ContainerDied","Data":"b3e56595afdbf4fa498f491dd5f6def1c362abf5289c49fe76865479c0af3207"} Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.844456 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3e83310b-1692-4be3-83ba-8278d0dc31e2","Type":"ContainerStarted","Data":"24707d6a99b3b11a63ec5f72981bca28ab0cd6e7ba06487e5749e6bc23addcf3"} Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.845371 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.931099 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.931077161 podStartE2EDuration="36.931077161s" podCreationTimestamp="2025-12-03 11:16:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:17:15.92425786 +0000 UTC m=+1412.387314025" watchObservedRunningTime="2025-12-03 11:17:15.931077161 +0000 UTC m=+1412.394133296" Dec 03 11:17:15 crc kubenswrapper[4646]: I1203 11:17:15.991632 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w"] Dec 03 11:17:16 crc kubenswrapper[4646]: I1203 11:17:16.859880 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" event={"ID":"5afc0cac-1a45-4f71-94e6-241231108859","Type":"ContainerStarted","Data":"1ef6f35655ea3a829a2160de58c82ee6ab4f41c62313f0d075c57a976b321834"} Dec 03 11:17:16 crc kubenswrapper[4646]: I1203 11:17:16.863062 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"16b4b1d1-a293-4042-b52b-dc8a10e07536","Type":"ContainerStarted","Data":"7b7fb1bbe09fe111ff0db42e13f1614413e491826c75521a2f581541e24fbbbc"} Dec 03 11:17:16 crc kubenswrapper[4646]: I1203 11:17:16.863273 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:17:16 crc kubenswrapper[4646]: I1203 11:17:16.899305 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.899284469 podStartE2EDuration="36.899284469s" podCreationTimestamp="2025-12-03 11:16:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 11:17:16.886898362 +0000 UTC m=+1413.349954527" watchObservedRunningTime="2025-12-03 11:17:16.899284469 +0000 UTC m=+1413.362340604" Dec 03 11:17:25 crc kubenswrapper[4646]: I1203 11:17:25.964729 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:17:25 crc kubenswrapper[4646]: I1203 11:17:25.965226 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:17:30 crc kubenswrapper[4646]: I1203 11:17:30.440588 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 11:17:30 crc kubenswrapper[4646]: I1203 11:17:30.615707 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" event={"ID":"5afc0cac-1a45-4f71-94e6-241231108859","Type":"ContainerStarted","Data":"f45cddaf20c28b45cb50be2cf58b0b3d60c8384312ad0b98c490a1da3424a49f"} Dec 03 11:17:30 crc kubenswrapper[4646]: I1203 11:17:30.650059 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" podStartSLOduration=1.657147697 podStartE2EDuration="15.650041417s" podCreationTimestamp="2025-12-03 11:17:15 +0000 UTC" firstStartedPulling="2025-12-03 11:17:16.006711519 +0000 UTC m=+1412.469767654" lastFinishedPulling="2025-12-03 11:17:29.999605239 +0000 UTC m=+1426.462661374" observedRunningTime="2025-12-03 11:17:30.646444365 +0000 UTC m=+1427.109500500" watchObservedRunningTime="2025-12-03 11:17:30.650041417 +0000 UTC m=+1427.113097552" Dec 03 11:17:31 crc kubenswrapper[4646]: I1203 11:17:31.182606 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 11:17:55 crc kubenswrapper[4646]: I1203 11:17:55.964735 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:17:55 crc kubenswrapper[4646]: I1203 11:17:55.965295 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:17:58 crc kubenswrapper[4646]: I1203 11:17:58.851490 4646 generic.go:334] "Generic (PLEG): container finished" podID="5afc0cac-1a45-4f71-94e6-241231108859" containerID="f45cddaf20c28b45cb50be2cf58b0b3d60c8384312ad0b98c490a1da3424a49f" exitCode=0 Dec 03 11:17:58 crc kubenswrapper[4646]: I1203 11:17:58.851566 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" event={"ID":"5afc0cac-1a45-4f71-94e6-241231108859","Type":"ContainerDied","Data":"f45cddaf20c28b45cb50be2cf58b0b3d60c8384312ad0b98c490a1da3424a49f"} Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.333850 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.513713 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnknp\" (UniqueName: \"kubernetes.io/projected/5afc0cac-1a45-4f71-94e6-241231108859-kube-api-access-rnknp\") pod \"5afc0cac-1a45-4f71-94e6-241231108859\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.513918 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-ssh-key\") pod \"5afc0cac-1a45-4f71-94e6-241231108859\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.514026 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-inventory\") pod \"5afc0cac-1a45-4f71-94e6-241231108859\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.514089 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-repo-setup-combined-ca-bundle\") pod \"5afc0cac-1a45-4f71-94e6-241231108859\" (UID: \"5afc0cac-1a45-4f71-94e6-241231108859\") " Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.525151 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5afc0cac-1a45-4f71-94e6-241231108859-kube-api-access-rnknp" (OuterVolumeSpecName: "kube-api-access-rnknp") pod "5afc0cac-1a45-4f71-94e6-241231108859" (UID: "5afc0cac-1a45-4f71-94e6-241231108859"). InnerVolumeSpecName "kube-api-access-rnknp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.530649 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "5afc0cac-1a45-4f71-94e6-241231108859" (UID: "5afc0cac-1a45-4f71-94e6-241231108859"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.546180 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-inventory" (OuterVolumeSpecName: "inventory") pod "5afc0cac-1a45-4f71-94e6-241231108859" (UID: "5afc0cac-1a45-4f71-94e6-241231108859"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.552052 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5afc0cac-1a45-4f71-94e6-241231108859" (UID: "5afc0cac-1a45-4f71-94e6-241231108859"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.616434 4646 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.616490 4646 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.616509 4646 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5afc0cac-1a45-4f71-94e6-241231108859-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.616530 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnknp\" (UniqueName: \"kubernetes.io/projected/5afc0cac-1a45-4f71-94e6-241231108859-kube-api-access-rnknp\") on node \"crc\" DevicePath \"\"" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.873902 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" event={"ID":"5afc0cac-1a45-4f71-94e6-241231108859","Type":"ContainerDied","Data":"1ef6f35655ea3a829a2160de58c82ee6ab4f41c62313f0d075c57a976b321834"} Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.873952 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.873962 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ef6f35655ea3a829a2160de58c82ee6ab4f41c62313f0d075c57a976b321834" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.964101 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5"] Dec 03 11:18:00 crc kubenswrapper[4646]: E1203 11:18:00.964771 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5afc0cac-1a45-4f71-94e6-241231108859" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.964858 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="5afc0cac-1a45-4f71-94e6-241231108859" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.965135 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="5afc0cac-1a45-4f71-94e6-241231108859" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.965834 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.969956 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.970273 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.970521 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.970880 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bfp25" Dec 03 11:18:00 crc kubenswrapper[4646]: I1203 11:18:00.974469 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5"] Dec 03 11:18:01 crc kubenswrapper[4646]: I1203 11:18:01.125037 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:18:01 crc kubenswrapper[4646]: I1203 11:18:01.125396 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:18:01 crc kubenswrapper[4646]: I1203 11:18:01.125427 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:18:01 crc kubenswrapper[4646]: I1203 11:18:01.125466 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r49pt\" (UniqueName: \"kubernetes.io/projected/d6fdb595-c620-4fd6-9a04-5acc542f158d-kube-api-access-r49pt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:18:01 crc kubenswrapper[4646]: I1203 11:18:01.227208 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:18:01 crc kubenswrapper[4646]: I1203 11:18:01.227530 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:18:01 crc kubenswrapper[4646]: I1203 11:18:01.227638 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:18:01 crc kubenswrapper[4646]: I1203 11:18:01.227818 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r49pt\" (UniqueName: \"kubernetes.io/projected/d6fdb595-c620-4fd6-9a04-5acc542f158d-kube-api-access-r49pt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:18:01 crc kubenswrapper[4646]: I1203 11:18:01.233145 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:18:01 crc kubenswrapper[4646]: I1203 11:18:01.233351 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:18:01 crc kubenswrapper[4646]: I1203 11:18:01.233598 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:18:01 crc kubenswrapper[4646]: I1203 11:18:01.270627 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r49pt\" (UniqueName: \"kubernetes.io/projected/d6fdb595-c620-4fd6-9a04-5acc542f158d-kube-api-access-r49pt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:18:01 crc kubenswrapper[4646]: I1203 11:18:01.285520 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:18:01 crc kubenswrapper[4646]: I1203 11:18:01.742367 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5"] Dec 03 11:18:01 crc kubenswrapper[4646]: I1203 11:18:01.882880 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" event={"ID":"d6fdb595-c620-4fd6-9a04-5acc542f158d","Type":"ContainerStarted","Data":"5cc22b5e75571d3640a8265c5258921f789bdfc2a27ab251cf14f86deee855de"} Dec 03 11:18:02 crc kubenswrapper[4646]: I1203 11:18:02.895797 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" event={"ID":"d6fdb595-c620-4fd6-9a04-5acc542f158d","Type":"ContainerStarted","Data":"d84367bfec6f4eb9761c26156f7bd16dead6b81e4e494201a62adb547ee90fca"} Dec 03 11:18:02 crc kubenswrapper[4646]: I1203 11:18:02.919756 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" podStartSLOduration=2.371711261 podStartE2EDuration="2.919730546s" podCreationTimestamp="2025-12-03 11:18:00 +0000 UTC" firstStartedPulling="2025-12-03 11:18:01.744447352 +0000 UTC m=+1458.207503487" lastFinishedPulling="2025-12-03 11:18:02.292466647 +0000 UTC m=+1458.755522772" observedRunningTime="2025-12-03 11:18:02.913008354 +0000 UTC m=+1459.376064529" watchObservedRunningTime="2025-12-03 11:18:02.919730546 +0000 UTC m=+1459.382786681" Dec 03 11:18:25 crc kubenswrapper[4646]: I1203 11:18:25.964815 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:18:25 crc kubenswrapper[4646]: I1203 11:18:25.965418 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:18:25 crc kubenswrapper[4646]: I1203 11:18:25.965465 4646 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 11:18:25 crc kubenswrapper[4646]: I1203 11:18:25.966461 4646 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1e85145461d5955667430f06bb56cf842506506ca5bbed595179c9c9f545d13b"} pod="openshift-machine-config-operator/machine-config-daemon-pndzt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 11:18:25 crc kubenswrapper[4646]: I1203 11:18:25.966540 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" containerID="cri-o://1e85145461d5955667430f06bb56cf842506506ca5bbed595179c9c9f545d13b" gracePeriod=600 Dec 03 11:18:26 crc kubenswrapper[4646]: I1203 11:18:26.110940 4646 generic.go:334] "Generic (PLEG): container finished" podID="9695d947-fc43-4625-86af-a1a3e177bb27" containerID="1e85145461d5955667430f06bb56cf842506506ca5bbed595179c9c9f545d13b" exitCode=0 Dec 03 11:18:26 crc kubenswrapper[4646]: I1203 11:18:26.110987 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerDied","Data":"1e85145461d5955667430f06bb56cf842506506ca5bbed595179c9c9f545d13b"} Dec 03 11:18:26 crc kubenswrapper[4646]: I1203 11:18:26.111044 4646 scope.go:117] "RemoveContainer" containerID="dfd3243f7f8d357efda5aacf86cad097d4462b47d7d216b50c99ea94ea34af46" Dec 03 11:18:27 crc kubenswrapper[4646]: I1203 11:18:27.120323 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerStarted","Data":"1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e"} Dec 03 11:19:03 crc kubenswrapper[4646]: I1203 11:19:03.702824 4646 scope.go:117] "RemoveContainer" containerID="6462075ca2426030733798995ebbc6119432ea1d42417a8a33ba051445c2bda9" Dec 03 11:19:03 crc kubenswrapper[4646]: I1203 11:19:03.743195 4646 scope.go:117] "RemoveContainer" containerID="6a79067ebbc0d9ee89292863e21922a5321ab9795bdca740f490c970db84a714" Dec 03 11:20:00 crc kubenswrapper[4646]: I1203 11:20:00.177682 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rnc9l"] Dec 03 11:20:00 crc kubenswrapper[4646]: I1203 11:20:00.180833 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:00 crc kubenswrapper[4646]: I1203 11:20:00.185554 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rnc9l"] Dec 03 11:20:00 crc kubenswrapper[4646]: I1203 11:20:00.350914 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-catalog-content\") pod \"community-operators-rnc9l\" (UID: \"1f9da2ed-bcf3-4c81-a11b-3966579c91ed\") " pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:00 crc kubenswrapper[4646]: I1203 11:20:00.351143 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-utilities\") pod \"community-operators-rnc9l\" (UID: \"1f9da2ed-bcf3-4c81-a11b-3966579c91ed\") " pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:00 crc kubenswrapper[4646]: I1203 11:20:00.351244 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdpbg\" (UniqueName: \"kubernetes.io/projected/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-kube-api-access-mdpbg\") pod \"community-operators-rnc9l\" (UID: \"1f9da2ed-bcf3-4c81-a11b-3966579c91ed\") " pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:00 crc kubenswrapper[4646]: I1203 11:20:00.453457 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-catalog-content\") pod \"community-operators-rnc9l\" (UID: \"1f9da2ed-bcf3-4c81-a11b-3966579c91ed\") " pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:00 crc kubenswrapper[4646]: I1203 11:20:00.453552 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-utilities\") pod \"community-operators-rnc9l\" (UID: \"1f9da2ed-bcf3-4c81-a11b-3966579c91ed\") " pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:00 crc kubenswrapper[4646]: I1203 11:20:00.453590 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdpbg\" (UniqueName: \"kubernetes.io/projected/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-kube-api-access-mdpbg\") pod \"community-operators-rnc9l\" (UID: \"1f9da2ed-bcf3-4c81-a11b-3966579c91ed\") " pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:00 crc kubenswrapper[4646]: I1203 11:20:00.454211 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-utilities\") pod \"community-operators-rnc9l\" (UID: \"1f9da2ed-bcf3-4c81-a11b-3966579c91ed\") " pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:00 crc kubenswrapper[4646]: I1203 11:20:00.454225 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-catalog-content\") pod \"community-operators-rnc9l\" (UID: \"1f9da2ed-bcf3-4c81-a11b-3966579c91ed\") " pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:00 crc kubenswrapper[4646]: I1203 11:20:00.476157 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdpbg\" (UniqueName: \"kubernetes.io/projected/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-kube-api-access-mdpbg\") pod \"community-operators-rnc9l\" (UID: \"1f9da2ed-bcf3-4c81-a11b-3966579c91ed\") " pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:00 crc kubenswrapper[4646]: I1203 11:20:00.504307 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:00 crc kubenswrapper[4646]: I1203 11:20:00.985833 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rnc9l"] Dec 03 11:20:01 crc kubenswrapper[4646]: I1203 11:20:01.057754 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnc9l" event={"ID":"1f9da2ed-bcf3-4c81-a11b-3966579c91ed","Type":"ContainerStarted","Data":"fcf3a67b3f7d67dfa50a5f6ef458cd54781b641dcab5f4df99afe334e706fa0a"} Dec 03 11:20:02 crc kubenswrapper[4646]: I1203 11:20:02.081271 4646 generic.go:334] "Generic (PLEG): container finished" podID="1f9da2ed-bcf3-4c81-a11b-3966579c91ed" containerID="5f6059ef94d139a2acafae369f69a593f62c8b6548b23c8e2e7f37b3283130f8" exitCode=0 Dec 03 11:20:02 crc kubenswrapper[4646]: I1203 11:20:02.081357 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnc9l" event={"ID":"1f9da2ed-bcf3-4c81-a11b-3966579c91ed","Type":"ContainerDied","Data":"5f6059ef94d139a2acafae369f69a593f62c8b6548b23c8e2e7f37b3283130f8"} Dec 03 11:20:02 crc kubenswrapper[4646]: I1203 11:20:02.085453 4646 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 11:20:03 crc kubenswrapper[4646]: I1203 11:20:03.091247 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnc9l" event={"ID":"1f9da2ed-bcf3-4c81-a11b-3966579c91ed","Type":"ContainerStarted","Data":"25bb4469886e7eac85df3907ded85964460c02d8a7955abd321e45b31aff8234"} Dec 03 11:20:05 crc kubenswrapper[4646]: I1203 11:20:05.110074 4646 generic.go:334] "Generic (PLEG): container finished" podID="1f9da2ed-bcf3-4c81-a11b-3966579c91ed" containerID="25bb4469886e7eac85df3907ded85964460c02d8a7955abd321e45b31aff8234" exitCode=0 Dec 03 11:20:05 crc kubenswrapper[4646]: I1203 11:20:05.110528 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnc9l" event={"ID":"1f9da2ed-bcf3-4c81-a11b-3966579c91ed","Type":"ContainerDied","Data":"25bb4469886e7eac85df3907ded85964460c02d8a7955abd321e45b31aff8234"} Dec 03 11:20:06 crc kubenswrapper[4646]: I1203 11:20:06.132995 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnc9l" event={"ID":"1f9da2ed-bcf3-4c81-a11b-3966579c91ed","Type":"ContainerStarted","Data":"e1531ea75b79ec746bdad6f297a038ec93539a8c03c0ee5cdf2cd52ed3143f41"} Dec 03 11:20:06 crc kubenswrapper[4646]: I1203 11:20:06.157567 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rnc9l" podStartSLOduration=2.695475239 podStartE2EDuration="6.15754624s" podCreationTimestamp="2025-12-03 11:20:00 +0000 UTC" firstStartedPulling="2025-12-03 11:20:02.084948407 +0000 UTC m=+1578.548004542" lastFinishedPulling="2025-12-03 11:20:05.547019398 +0000 UTC m=+1582.010075543" observedRunningTime="2025-12-03 11:20:06.152284152 +0000 UTC m=+1582.615340297" watchObservedRunningTime="2025-12-03 11:20:06.15754624 +0000 UTC m=+1582.620602375" Dec 03 11:20:09 crc kubenswrapper[4646]: I1203 11:20:09.055965 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dgsfw"] Dec 03 11:20:09 crc kubenswrapper[4646]: I1203 11:20:09.058879 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:09 crc kubenswrapper[4646]: I1203 11:20:09.068441 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dgsfw"] Dec 03 11:20:09 crc kubenswrapper[4646]: I1203 11:20:09.115180 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-catalog-content\") pod \"redhat-marketplace-dgsfw\" (UID: \"debb41a6-c8b9-43c8-be2a-5d94bbe213b2\") " pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:09 crc kubenswrapper[4646]: I1203 11:20:09.115280 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-utilities\") pod \"redhat-marketplace-dgsfw\" (UID: \"debb41a6-c8b9-43c8-be2a-5d94bbe213b2\") " pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:09 crc kubenswrapper[4646]: I1203 11:20:09.115328 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w66k\" (UniqueName: \"kubernetes.io/projected/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-kube-api-access-2w66k\") pod \"redhat-marketplace-dgsfw\" (UID: \"debb41a6-c8b9-43c8-be2a-5d94bbe213b2\") " pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:09 crc kubenswrapper[4646]: I1203 11:20:09.217055 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-catalog-content\") pod \"redhat-marketplace-dgsfw\" (UID: \"debb41a6-c8b9-43c8-be2a-5d94bbe213b2\") " pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:09 crc kubenswrapper[4646]: I1203 11:20:09.217137 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-utilities\") pod \"redhat-marketplace-dgsfw\" (UID: \"debb41a6-c8b9-43c8-be2a-5d94bbe213b2\") " pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:09 crc kubenswrapper[4646]: I1203 11:20:09.217170 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w66k\" (UniqueName: \"kubernetes.io/projected/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-kube-api-access-2w66k\") pod \"redhat-marketplace-dgsfw\" (UID: \"debb41a6-c8b9-43c8-be2a-5d94bbe213b2\") " pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:09 crc kubenswrapper[4646]: I1203 11:20:09.217972 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-catalog-content\") pod \"redhat-marketplace-dgsfw\" (UID: \"debb41a6-c8b9-43c8-be2a-5d94bbe213b2\") " pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:09 crc kubenswrapper[4646]: I1203 11:20:09.218047 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-utilities\") pod \"redhat-marketplace-dgsfw\" (UID: \"debb41a6-c8b9-43c8-be2a-5d94bbe213b2\") " pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:09 crc kubenswrapper[4646]: I1203 11:20:09.236227 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w66k\" (UniqueName: \"kubernetes.io/projected/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-kube-api-access-2w66k\") pod \"redhat-marketplace-dgsfw\" (UID: \"debb41a6-c8b9-43c8-be2a-5d94bbe213b2\") " pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:09 crc kubenswrapper[4646]: I1203 11:20:09.381977 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:09 crc kubenswrapper[4646]: I1203 11:20:09.859129 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dgsfw"] Dec 03 11:20:10 crc kubenswrapper[4646]: I1203 11:20:10.165683 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgsfw" event={"ID":"debb41a6-c8b9-43c8-be2a-5d94bbe213b2","Type":"ContainerStarted","Data":"feb6aa10315f5945054d24f961ff67cfd58045e8147424e9f99afa4b05214b3f"} Dec 03 11:20:10 crc kubenswrapper[4646]: I1203 11:20:10.505152 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:10 crc kubenswrapper[4646]: I1203 11:20:10.505211 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:10 crc kubenswrapper[4646]: I1203 11:20:10.555661 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:11 crc kubenswrapper[4646]: I1203 11:20:11.176604 4646 generic.go:334] "Generic (PLEG): container finished" podID="debb41a6-c8b9-43c8-be2a-5d94bbe213b2" containerID="1e26e941135ec630633e223d072abe2f1c67bcda2596836c477b047a94683fb4" exitCode=0 Dec 03 11:20:11 crc kubenswrapper[4646]: I1203 11:20:11.178196 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgsfw" event={"ID":"debb41a6-c8b9-43c8-be2a-5d94bbe213b2","Type":"ContainerDied","Data":"1e26e941135ec630633e223d072abe2f1c67bcda2596836c477b047a94683fb4"} Dec 03 11:20:11 crc kubenswrapper[4646]: I1203 11:20:11.248224 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:12 crc kubenswrapper[4646]: I1203 11:20:12.846887 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rnc9l"] Dec 03 11:20:13 crc kubenswrapper[4646]: I1203 11:20:13.199800 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgsfw" event={"ID":"debb41a6-c8b9-43c8-be2a-5d94bbe213b2","Type":"ContainerStarted","Data":"e482dd8d4e6dbfe24a08503251f79ce40edcfd6107e6f8805d7ac44972caa40e"} Dec 03 11:20:13 crc kubenswrapper[4646]: I1203 11:20:13.199971 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rnc9l" podUID="1f9da2ed-bcf3-4c81-a11b-3966579c91ed" containerName="registry-server" containerID="cri-o://e1531ea75b79ec746bdad6f297a038ec93539a8c03c0ee5cdf2cd52ed3143f41" gracePeriod=2 Dec 03 11:20:15 crc kubenswrapper[4646]: I1203 11:20:15.222290 4646 generic.go:334] "Generic (PLEG): container finished" podID="debb41a6-c8b9-43c8-be2a-5d94bbe213b2" containerID="e482dd8d4e6dbfe24a08503251f79ce40edcfd6107e6f8805d7ac44972caa40e" exitCode=0 Dec 03 11:20:15 crc kubenswrapper[4646]: I1203 11:20:15.222391 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgsfw" event={"ID":"debb41a6-c8b9-43c8-be2a-5d94bbe213b2","Type":"ContainerDied","Data":"e482dd8d4e6dbfe24a08503251f79ce40edcfd6107e6f8805d7ac44972caa40e"} Dec 03 11:20:15 crc kubenswrapper[4646]: I1203 11:20:15.232166 4646 generic.go:334] "Generic (PLEG): container finished" podID="1f9da2ed-bcf3-4c81-a11b-3966579c91ed" containerID="e1531ea75b79ec746bdad6f297a038ec93539a8c03c0ee5cdf2cd52ed3143f41" exitCode=0 Dec 03 11:20:15 crc kubenswrapper[4646]: I1203 11:20:15.232217 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnc9l" event={"ID":"1f9da2ed-bcf3-4c81-a11b-3966579c91ed","Type":"ContainerDied","Data":"e1531ea75b79ec746bdad6f297a038ec93539a8c03c0ee5cdf2cd52ed3143f41"} Dec 03 11:20:15 crc kubenswrapper[4646]: I1203 11:20:15.476129 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:15 crc kubenswrapper[4646]: I1203 11:20:15.551165 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdpbg\" (UniqueName: \"kubernetes.io/projected/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-kube-api-access-mdpbg\") pod \"1f9da2ed-bcf3-4c81-a11b-3966579c91ed\" (UID: \"1f9da2ed-bcf3-4c81-a11b-3966579c91ed\") " Dec 03 11:20:15 crc kubenswrapper[4646]: I1203 11:20:15.551245 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-catalog-content\") pod \"1f9da2ed-bcf3-4c81-a11b-3966579c91ed\" (UID: \"1f9da2ed-bcf3-4c81-a11b-3966579c91ed\") " Dec 03 11:20:15 crc kubenswrapper[4646]: I1203 11:20:15.551309 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-utilities\") pod \"1f9da2ed-bcf3-4c81-a11b-3966579c91ed\" (UID: \"1f9da2ed-bcf3-4c81-a11b-3966579c91ed\") " Dec 03 11:20:15 crc kubenswrapper[4646]: I1203 11:20:15.552303 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-utilities" (OuterVolumeSpecName: "utilities") pod "1f9da2ed-bcf3-4c81-a11b-3966579c91ed" (UID: "1f9da2ed-bcf3-4c81-a11b-3966579c91ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:20:15 crc kubenswrapper[4646]: I1203 11:20:15.557021 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-kube-api-access-mdpbg" (OuterVolumeSpecName: "kube-api-access-mdpbg") pod "1f9da2ed-bcf3-4c81-a11b-3966579c91ed" (UID: "1f9da2ed-bcf3-4c81-a11b-3966579c91ed"). InnerVolumeSpecName "kube-api-access-mdpbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:20:15 crc kubenswrapper[4646]: I1203 11:20:15.606367 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f9da2ed-bcf3-4c81-a11b-3966579c91ed" (UID: "1f9da2ed-bcf3-4c81-a11b-3966579c91ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:20:15 crc kubenswrapper[4646]: I1203 11:20:15.653220 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 11:20:15 crc kubenswrapper[4646]: I1203 11:20:15.653500 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 11:20:15 crc kubenswrapper[4646]: I1203 11:20:15.653589 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdpbg\" (UniqueName: \"kubernetes.io/projected/1f9da2ed-bcf3-4c81-a11b-3966579c91ed-kube-api-access-mdpbg\") on node \"crc\" DevicePath \"\"" Dec 03 11:20:16 crc kubenswrapper[4646]: I1203 11:20:16.243968 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnc9l" event={"ID":"1f9da2ed-bcf3-4c81-a11b-3966579c91ed","Type":"ContainerDied","Data":"fcf3a67b3f7d67dfa50a5f6ef458cd54781b641dcab5f4df99afe334e706fa0a"} Dec 03 11:20:16 crc kubenswrapper[4646]: I1203 11:20:16.243998 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnc9l" Dec 03 11:20:16 crc kubenswrapper[4646]: I1203 11:20:16.244027 4646 scope.go:117] "RemoveContainer" containerID="e1531ea75b79ec746bdad6f297a038ec93539a8c03c0ee5cdf2cd52ed3143f41" Dec 03 11:20:16 crc kubenswrapper[4646]: I1203 11:20:16.249185 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgsfw" event={"ID":"debb41a6-c8b9-43c8-be2a-5d94bbe213b2","Type":"ContainerStarted","Data":"69296f9c70a65b23cfc92433a4cffc378b5813d555980df6b8e739325a49575c"} Dec 03 11:20:16 crc kubenswrapper[4646]: I1203 11:20:16.270018 4646 scope.go:117] "RemoveContainer" containerID="25bb4469886e7eac85df3907ded85964460c02d8a7955abd321e45b31aff8234" Dec 03 11:20:16 crc kubenswrapper[4646]: I1203 11:20:16.271313 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rnc9l"] Dec 03 11:20:16 crc kubenswrapper[4646]: I1203 11:20:16.280521 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rnc9l"] Dec 03 11:20:16 crc kubenswrapper[4646]: I1203 11:20:16.297198 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dgsfw" podStartSLOduration=2.811101575 podStartE2EDuration="7.29717964s" podCreationTimestamp="2025-12-03 11:20:09 +0000 UTC" firstStartedPulling="2025-12-03 11:20:11.179710883 +0000 UTC m=+1587.642767018" lastFinishedPulling="2025-12-03 11:20:15.665788948 +0000 UTC m=+1592.128845083" observedRunningTime="2025-12-03 11:20:16.292710355 +0000 UTC m=+1592.755766500" watchObservedRunningTime="2025-12-03 11:20:16.29717964 +0000 UTC m=+1592.760235775" Dec 03 11:20:16 crc kubenswrapper[4646]: I1203 11:20:16.298607 4646 scope.go:117] "RemoveContainer" containerID="5f6059ef94d139a2acafae369f69a593f62c8b6548b23c8e2e7f37b3283130f8" Dec 03 11:20:17 crc kubenswrapper[4646]: I1203 11:20:17.860295 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f9da2ed-bcf3-4c81-a11b-3966579c91ed" path="/var/lib/kubelet/pods/1f9da2ed-bcf3-4c81-a11b-3966579c91ed/volumes" Dec 03 11:20:19 crc kubenswrapper[4646]: I1203 11:20:19.382797 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:19 crc kubenswrapper[4646]: I1203 11:20:19.382889 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:19 crc kubenswrapper[4646]: I1203 11:20:19.438804 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:20 crc kubenswrapper[4646]: I1203 11:20:20.331863 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:20 crc kubenswrapper[4646]: I1203 11:20:20.834055 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dgsfw"] Dec 03 11:20:22 crc kubenswrapper[4646]: I1203 11:20:22.303516 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dgsfw" podUID="debb41a6-c8b9-43c8-be2a-5d94bbe213b2" containerName="registry-server" containerID="cri-o://69296f9c70a65b23cfc92433a4cffc378b5813d555980df6b8e739325a49575c" gracePeriod=2 Dec 03 11:20:22 crc kubenswrapper[4646]: I1203 11:20:22.842667 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.014561 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-utilities\") pod \"debb41a6-c8b9-43c8-be2a-5d94bbe213b2\" (UID: \"debb41a6-c8b9-43c8-be2a-5d94bbe213b2\") " Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.014706 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w66k\" (UniqueName: \"kubernetes.io/projected/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-kube-api-access-2w66k\") pod \"debb41a6-c8b9-43c8-be2a-5d94bbe213b2\" (UID: \"debb41a6-c8b9-43c8-be2a-5d94bbe213b2\") " Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.014833 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-catalog-content\") pod \"debb41a6-c8b9-43c8-be2a-5d94bbe213b2\" (UID: \"debb41a6-c8b9-43c8-be2a-5d94bbe213b2\") " Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.016369 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-utilities" (OuterVolumeSpecName: "utilities") pod "debb41a6-c8b9-43c8-be2a-5d94bbe213b2" (UID: "debb41a6-c8b9-43c8-be2a-5d94bbe213b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.025742 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-kube-api-access-2w66k" (OuterVolumeSpecName: "kube-api-access-2w66k") pod "debb41a6-c8b9-43c8-be2a-5d94bbe213b2" (UID: "debb41a6-c8b9-43c8-be2a-5d94bbe213b2"). InnerVolumeSpecName "kube-api-access-2w66k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.041731 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "debb41a6-c8b9-43c8-be2a-5d94bbe213b2" (UID: "debb41a6-c8b9-43c8-be2a-5d94bbe213b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.117030 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.117064 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.117074 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w66k\" (UniqueName: \"kubernetes.io/projected/debb41a6-c8b9-43c8-be2a-5d94bbe213b2-kube-api-access-2w66k\") on node \"crc\" DevicePath \"\"" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.314943 4646 generic.go:334] "Generic (PLEG): container finished" podID="debb41a6-c8b9-43c8-be2a-5d94bbe213b2" containerID="69296f9c70a65b23cfc92433a4cffc378b5813d555980df6b8e739325a49575c" exitCode=0 Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.314991 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgsfw" event={"ID":"debb41a6-c8b9-43c8-be2a-5d94bbe213b2","Type":"ContainerDied","Data":"69296f9c70a65b23cfc92433a4cffc378b5813d555980df6b8e739325a49575c"} Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.315002 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dgsfw" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.315029 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgsfw" event={"ID":"debb41a6-c8b9-43c8-be2a-5d94bbe213b2","Type":"ContainerDied","Data":"feb6aa10315f5945054d24f961ff67cfd58045e8147424e9f99afa4b05214b3f"} Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.315047 4646 scope.go:117] "RemoveContainer" containerID="69296f9c70a65b23cfc92433a4cffc378b5813d555980df6b8e739325a49575c" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.355818 4646 scope.go:117] "RemoveContainer" containerID="e482dd8d4e6dbfe24a08503251f79ce40edcfd6107e6f8805d7ac44972caa40e" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.383518 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dgsfw"] Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.390001 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dgsfw"] Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.392298 4646 scope.go:117] "RemoveContainer" containerID="1e26e941135ec630633e223d072abe2f1c67bcda2596836c477b047a94683fb4" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.433274 4646 scope.go:117] "RemoveContainer" containerID="69296f9c70a65b23cfc92433a4cffc378b5813d555980df6b8e739325a49575c" Dec 03 11:20:23 crc kubenswrapper[4646]: E1203 11:20:23.434006 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69296f9c70a65b23cfc92433a4cffc378b5813d555980df6b8e739325a49575c\": container with ID starting with 69296f9c70a65b23cfc92433a4cffc378b5813d555980df6b8e739325a49575c not found: ID does not exist" containerID="69296f9c70a65b23cfc92433a4cffc378b5813d555980df6b8e739325a49575c" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.434082 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69296f9c70a65b23cfc92433a4cffc378b5813d555980df6b8e739325a49575c"} err="failed to get container status \"69296f9c70a65b23cfc92433a4cffc378b5813d555980df6b8e739325a49575c\": rpc error: code = NotFound desc = could not find container \"69296f9c70a65b23cfc92433a4cffc378b5813d555980df6b8e739325a49575c\": container with ID starting with 69296f9c70a65b23cfc92433a4cffc378b5813d555980df6b8e739325a49575c not found: ID does not exist" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.434115 4646 scope.go:117] "RemoveContainer" containerID="e482dd8d4e6dbfe24a08503251f79ce40edcfd6107e6f8805d7ac44972caa40e" Dec 03 11:20:23 crc kubenswrapper[4646]: E1203 11:20:23.434596 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e482dd8d4e6dbfe24a08503251f79ce40edcfd6107e6f8805d7ac44972caa40e\": container with ID starting with e482dd8d4e6dbfe24a08503251f79ce40edcfd6107e6f8805d7ac44972caa40e not found: ID does not exist" containerID="e482dd8d4e6dbfe24a08503251f79ce40edcfd6107e6f8805d7ac44972caa40e" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.434647 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e482dd8d4e6dbfe24a08503251f79ce40edcfd6107e6f8805d7ac44972caa40e"} err="failed to get container status \"e482dd8d4e6dbfe24a08503251f79ce40edcfd6107e6f8805d7ac44972caa40e\": rpc error: code = NotFound desc = could not find container \"e482dd8d4e6dbfe24a08503251f79ce40edcfd6107e6f8805d7ac44972caa40e\": container with ID starting with e482dd8d4e6dbfe24a08503251f79ce40edcfd6107e6f8805d7ac44972caa40e not found: ID does not exist" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.434684 4646 scope.go:117] "RemoveContainer" containerID="1e26e941135ec630633e223d072abe2f1c67bcda2596836c477b047a94683fb4" Dec 03 11:20:23 crc kubenswrapper[4646]: E1203 11:20:23.434998 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e26e941135ec630633e223d072abe2f1c67bcda2596836c477b047a94683fb4\": container with ID starting with 1e26e941135ec630633e223d072abe2f1c67bcda2596836c477b047a94683fb4 not found: ID does not exist" containerID="1e26e941135ec630633e223d072abe2f1c67bcda2596836c477b047a94683fb4" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.435020 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e26e941135ec630633e223d072abe2f1c67bcda2596836c477b047a94683fb4"} err="failed to get container status \"1e26e941135ec630633e223d072abe2f1c67bcda2596836c477b047a94683fb4\": rpc error: code = NotFound desc = could not find container \"1e26e941135ec630633e223d072abe2f1c67bcda2596836c477b047a94683fb4\": container with ID starting with 1e26e941135ec630633e223d072abe2f1c67bcda2596836c477b047a94683fb4 not found: ID does not exist" Dec 03 11:20:23 crc kubenswrapper[4646]: I1203 11:20:23.861704 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="debb41a6-c8b9-43c8-be2a-5d94bbe213b2" path="/var/lib/kubelet/pods/debb41a6-c8b9-43c8-be2a-5d94bbe213b2/volumes" Dec 03 11:20:55 crc kubenswrapper[4646]: I1203 11:20:55.964758 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:20:55 crc kubenswrapper[4646]: I1203 11:20:55.965325 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:21:25 crc kubenswrapper[4646]: I1203 11:21:25.964558 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:21:25 crc kubenswrapper[4646]: I1203 11:21:25.965108 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:21:55 crc kubenswrapper[4646]: I1203 11:21:55.964404 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:21:55 crc kubenswrapper[4646]: I1203 11:21:55.966251 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:21:55 crc kubenswrapper[4646]: I1203 11:21:55.966471 4646 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 11:21:55 crc kubenswrapper[4646]: I1203 11:21:55.967447 4646 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e"} pod="openshift-machine-config-operator/machine-config-daemon-pndzt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 11:21:55 crc kubenswrapper[4646]: I1203 11:21:55.967623 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" containerID="cri-o://1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" gracePeriod=600 Dec 03 11:21:56 crc kubenswrapper[4646]: E1203 11:21:56.116831 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:21:56 crc kubenswrapper[4646]: I1203 11:21:56.216201 4646 generic.go:334] "Generic (PLEG): container finished" podID="9695d947-fc43-4625-86af-a1a3e177bb27" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" exitCode=0 Dec 03 11:21:56 crc kubenswrapper[4646]: I1203 11:21:56.216219 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerDied","Data":"1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e"} Dec 03 11:21:56 crc kubenswrapper[4646]: I1203 11:21:56.216319 4646 scope.go:117] "RemoveContainer" containerID="1e85145461d5955667430f06bb56cf842506506ca5bbed595179c9c9f545d13b" Dec 03 11:21:56 crc kubenswrapper[4646]: I1203 11:21:56.216757 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:21:56 crc kubenswrapper[4646]: E1203 11:21:56.217184 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:21:58 crc kubenswrapper[4646]: I1203 11:21:58.249247 4646 generic.go:334] "Generic (PLEG): container finished" podID="d6fdb595-c620-4fd6-9a04-5acc542f158d" containerID="d84367bfec6f4eb9761c26156f7bd16dead6b81e4e494201a62adb547ee90fca" exitCode=0 Dec 03 11:21:58 crc kubenswrapper[4646]: I1203 11:21:58.249432 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" event={"ID":"d6fdb595-c620-4fd6-9a04-5acc542f158d","Type":"ContainerDied","Data":"d84367bfec6f4eb9761c26156f7bd16dead6b81e4e494201a62adb547ee90fca"} Dec 03 11:21:59 crc kubenswrapper[4646]: I1203 11:21:59.650607 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:21:59 crc kubenswrapper[4646]: I1203 11:21:59.682637 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-ssh-key\") pod \"d6fdb595-c620-4fd6-9a04-5acc542f158d\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " Dec 03 11:21:59 crc kubenswrapper[4646]: I1203 11:21:59.682762 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-inventory\") pod \"d6fdb595-c620-4fd6-9a04-5acc542f158d\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " Dec 03 11:21:59 crc kubenswrapper[4646]: I1203 11:21:59.682902 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-bootstrap-combined-ca-bundle\") pod \"d6fdb595-c620-4fd6-9a04-5acc542f158d\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " Dec 03 11:21:59 crc kubenswrapper[4646]: I1203 11:21:59.682949 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r49pt\" (UniqueName: \"kubernetes.io/projected/d6fdb595-c620-4fd6-9a04-5acc542f158d-kube-api-access-r49pt\") pod \"d6fdb595-c620-4fd6-9a04-5acc542f158d\" (UID: \"d6fdb595-c620-4fd6-9a04-5acc542f158d\") " Dec 03 11:21:59 crc kubenswrapper[4646]: I1203 11:21:59.693652 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d6fdb595-c620-4fd6-9a04-5acc542f158d" (UID: "d6fdb595-c620-4fd6-9a04-5acc542f158d"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:21:59 crc kubenswrapper[4646]: I1203 11:21:59.693739 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6fdb595-c620-4fd6-9a04-5acc542f158d-kube-api-access-r49pt" (OuterVolumeSpecName: "kube-api-access-r49pt") pod "d6fdb595-c620-4fd6-9a04-5acc542f158d" (UID: "d6fdb595-c620-4fd6-9a04-5acc542f158d"). InnerVolumeSpecName "kube-api-access-r49pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:21:59 crc kubenswrapper[4646]: I1203 11:21:59.725444 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d6fdb595-c620-4fd6-9a04-5acc542f158d" (UID: "d6fdb595-c620-4fd6-9a04-5acc542f158d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:21:59 crc kubenswrapper[4646]: I1203 11:21:59.728373 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-inventory" (OuterVolumeSpecName: "inventory") pod "d6fdb595-c620-4fd6-9a04-5acc542f158d" (UID: "d6fdb595-c620-4fd6-9a04-5acc542f158d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:21:59 crc kubenswrapper[4646]: I1203 11:21:59.785027 4646 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 11:21:59 crc kubenswrapper[4646]: I1203 11:21:59.785380 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r49pt\" (UniqueName: \"kubernetes.io/projected/d6fdb595-c620-4fd6-9a04-5acc542f158d-kube-api-access-r49pt\") on node \"crc\" DevicePath \"\"" Dec 03 11:21:59 crc kubenswrapper[4646]: I1203 11:21:59.785394 4646 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 11:21:59 crc kubenswrapper[4646]: I1203 11:21:59.785402 4646 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6fdb595-c620-4fd6-9a04-5acc542f158d-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.272525 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" event={"ID":"d6fdb595-c620-4fd6-9a04-5acc542f158d","Type":"ContainerDied","Data":"5cc22b5e75571d3640a8265c5258921f789bdfc2a27ab251cf14f86deee855de"} Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.273294 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cc22b5e75571d3640a8265c5258921f789bdfc2a27ab251cf14f86deee855de" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.272600 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.408016 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm"] Dec 03 11:22:00 crc kubenswrapper[4646]: E1203 11:22:00.408483 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="debb41a6-c8b9-43c8-be2a-5d94bbe213b2" containerName="extract-content" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.408503 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="debb41a6-c8b9-43c8-be2a-5d94bbe213b2" containerName="extract-content" Dec 03 11:22:00 crc kubenswrapper[4646]: E1203 11:22:00.408522 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6fdb595-c620-4fd6-9a04-5acc542f158d" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.408530 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6fdb595-c620-4fd6-9a04-5acc542f158d" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 11:22:00 crc kubenswrapper[4646]: E1203 11:22:00.408558 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9da2ed-bcf3-4c81-a11b-3966579c91ed" containerName="extract-content" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.408565 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9da2ed-bcf3-4c81-a11b-3966579c91ed" containerName="extract-content" Dec 03 11:22:00 crc kubenswrapper[4646]: E1203 11:22:00.408579 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9da2ed-bcf3-4c81-a11b-3966579c91ed" containerName="extract-utilities" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.408587 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9da2ed-bcf3-4c81-a11b-3966579c91ed" containerName="extract-utilities" Dec 03 11:22:00 crc kubenswrapper[4646]: E1203 11:22:00.408625 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="debb41a6-c8b9-43c8-be2a-5d94bbe213b2" containerName="extract-utilities" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.408633 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="debb41a6-c8b9-43c8-be2a-5d94bbe213b2" containerName="extract-utilities" Dec 03 11:22:00 crc kubenswrapper[4646]: E1203 11:22:00.408649 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="debb41a6-c8b9-43c8-be2a-5d94bbe213b2" containerName="registry-server" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.408656 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="debb41a6-c8b9-43c8-be2a-5d94bbe213b2" containerName="registry-server" Dec 03 11:22:00 crc kubenswrapper[4646]: E1203 11:22:00.408671 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9da2ed-bcf3-4c81-a11b-3966579c91ed" containerName="registry-server" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.408678 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9da2ed-bcf3-4c81-a11b-3966579c91ed" containerName="registry-server" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.408854 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="debb41a6-c8b9-43c8-be2a-5d94bbe213b2" containerName="registry-server" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.408870 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f9da2ed-bcf3-4c81-a11b-3966579c91ed" containerName="registry-server" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.408880 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6fdb595-c620-4fd6-9a04-5acc542f158d" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.409588 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.415067 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bfp25" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.415165 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.417395 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.419832 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.420192 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm"] Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.503999 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msqgs\" (UniqueName: \"kubernetes.io/projected/6185a07d-c02f-42b5-b060-c5317850e67b-kube-api-access-msqgs\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm\" (UID: \"6185a07d-c02f-42b5-b060-c5317850e67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.504062 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6185a07d-c02f-42b5-b060-c5317850e67b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm\" (UID: \"6185a07d-c02f-42b5-b060-c5317850e67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.504234 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6185a07d-c02f-42b5-b060-c5317850e67b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm\" (UID: \"6185a07d-c02f-42b5-b060-c5317850e67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.605445 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msqgs\" (UniqueName: \"kubernetes.io/projected/6185a07d-c02f-42b5-b060-c5317850e67b-kube-api-access-msqgs\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm\" (UID: \"6185a07d-c02f-42b5-b060-c5317850e67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.605516 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6185a07d-c02f-42b5-b060-c5317850e67b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm\" (UID: \"6185a07d-c02f-42b5-b060-c5317850e67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.605620 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6185a07d-c02f-42b5-b060-c5317850e67b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm\" (UID: \"6185a07d-c02f-42b5-b060-c5317850e67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.609992 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6185a07d-c02f-42b5-b060-c5317850e67b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm\" (UID: \"6185a07d-c02f-42b5-b060-c5317850e67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.610858 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6185a07d-c02f-42b5-b060-c5317850e67b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm\" (UID: \"6185a07d-c02f-42b5-b060-c5317850e67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.624047 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msqgs\" (UniqueName: \"kubernetes.io/projected/6185a07d-c02f-42b5-b060-c5317850e67b-kube-api-access-msqgs\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm\" (UID: \"6185a07d-c02f-42b5-b060-c5317850e67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" Dec 03 11:22:00 crc kubenswrapper[4646]: I1203 11:22:00.734268 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" Dec 03 11:22:01 crc kubenswrapper[4646]: I1203 11:22:01.304714 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm"] Dec 03 11:22:02 crc kubenswrapper[4646]: I1203 11:22:02.293557 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" event={"ID":"6185a07d-c02f-42b5-b060-c5317850e67b","Type":"ContainerStarted","Data":"5844d798323842ee5b6f6b2a53b8d2b5cd33c231dd845780b3e3cd6615c1e51e"} Dec 03 11:22:04 crc kubenswrapper[4646]: I1203 11:22:04.085926 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-xg9bs"] Dec 03 11:22:04 crc kubenswrapper[4646]: I1203 11:22:04.104559 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9826-account-create-update-9ldlv"] Dec 03 11:22:04 crc kubenswrapper[4646]: I1203 11:22:04.117296 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-26vjv"] Dec 03 11:22:04 crc kubenswrapper[4646]: I1203 11:22:04.126768 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bdf0-account-create-update-z84rm"] Dec 03 11:22:04 crc kubenswrapper[4646]: I1203 11:22:04.135618 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-6grfm"] Dec 03 11:22:04 crc kubenswrapper[4646]: I1203 11:22:04.144796 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-6grfm"] Dec 03 11:22:04 crc kubenswrapper[4646]: I1203 11:22:04.152783 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-11b8-account-create-update-khjqk"] Dec 03 11:22:04 crc kubenswrapper[4646]: I1203 11:22:04.163242 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-26vjv"] Dec 03 11:22:04 crc kubenswrapper[4646]: I1203 11:22:04.172403 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-9826-account-create-update-9ldlv"] Dec 03 11:22:04 crc kubenswrapper[4646]: I1203 11:22:04.181164 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-11b8-account-create-update-khjqk"] Dec 03 11:22:04 crc kubenswrapper[4646]: I1203 11:22:04.192895 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bdf0-account-create-update-z84rm"] Dec 03 11:22:04 crc kubenswrapper[4646]: I1203 11:22:04.201488 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-xg9bs"] Dec 03 11:22:04 crc kubenswrapper[4646]: I1203 11:22:04.343634 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" event={"ID":"6185a07d-c02f-42b5-b060-c5317850e67b","Type":"ContainerStarted","Data":"9845dcf8a51e9171a44b882eb3e2c6081706b8c7481958c4bcddb26045b28539"} Dec 03 11:22:05 crc kubenswrapper[4646]: I1203 11:22:05.871234 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22952a9c-263f-40d2-8f5f-54ef029814c8" path="/var/lib/kubelet/pods/22952a9c-263f-40d2-8f5f-54ef029814c8/volumes" Dec 03 11:22:05 crc kubenswrapper[4646]: I1203 11:22:05.872249 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="237d9726-11ed-4d72-8135-b8ba2f14813f" path="/var/lib/kubelet/pods/237d9726-11ed-4d72-8135-b8ba2f14813f/volumes" Dec 03 11:22:05 crc kubenswrapper[4646]: I1203 11:22:05.874749 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85bbeb75-a9c1-4812-916c-53bf850f1cb9" path="/var/lib/kubelet/pods/85bbeb75-a9c1-4812-916c-53bf850f1cb9/volumes" Dec 03 11:22:05 crc kubenswrapper[4646]: I1203 11:22:05.875796 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af4fb79d-1d31-4ae4-8769-ff0e6bae317f" path="/var/lib/kubelet/pods/af4fb79d-1d31-4ae4-8769-ff0e6bae317f/volumes" Dec 03 11:22:05 crc kubenswrapper[4646]: I1203 11:22:05.876569 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0660b75-6739-4a9e-8cb4-ff6b15b0080f" path="/var/lib/kubelet/pods/b0660b75-6739-4a9e-8cb4-ff6b15b0080f/volumes" Dec 03 11:22:05 crc kubenswrapper[4646]: I1203 11:22:05.877223 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b601410d-7e0e-48e7-b2cb-81126c9ed96b" path="/var/lib/kubelet/pods/b601410d-7e0e-48e7-b2cb-81126c9ed96b/volumes" Dec 03 11:22:06 crc kubenswrapper[4646]: I1203 11:22:06.848941 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:22:06 crc kubenswrapper[4646]: E1203 11:22:06.849472 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:22:20 crc kubenswrapper[4646]: I1203 11:22:20.849322 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:22:20 crc kubenswrapper[4646]: E1203 11:22:20.850279 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:22:28 crc kubenswrapper[4646]: I1203 11:22:28.084822 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" podStartSLOduration=25.446328196 podStartE2EDuration="28.084801429s" podCreationTimestamp="2025-12-03 11:22:00 +0000 UTC" firstStartedPulling="2025-12-03 11:22:01.311015742 +0000 UTC m=+1697.774071877" lastFinishedPulling="2025-12-03 11:22:03.949488975 +0000 UTC m=+1700.412545110" observedRunningTime="2025-12-03 11:22:04.362615514 +0000 UTC m=+1700.825671659" watchObservedRunningTime="2025-12-03 11:22:28.084801429 +0000 UTC m=+1724.547857564" Dec 03 11:22:28 crc kubenswrapper[4646]: I1203 11:22:28.094953 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-w557d"] Dec 03 11:22:28 crc kubenswrapper[4646]: I1203 11:22:28.106503 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-9dde-account-create-update-7ch8v"] Dec 03 11:22:28 crc kubenswrapper[4646]: I1203 11:22:28.117217 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-5129-account-create-update-drkql"] Dec 03 11:22:28 crc kubenswrapper[4646]: I1203 11:22:28.142364 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-d95tn"] Dec 03 11:22:28 crc kubenswrapper[4646]: I1203 11:22:28.153147 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-d2vp8"] Dec 03 11:22:28 crc kubenswrapper[4646]: I1203 11:22:28.161681 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-5129-account-create-update-drkql"] Dec 03 11:22:28 crc kubenswrapper[4646]: I1203 11:22:28.169879 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-e410-account-create-update-2nflx"] Dec 03 11:22:28 crc kubenswrapper[4646]: I1203 11:22:28.178131 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-9dde-account-create-update-7ch8v"] Dec 03 11:22:28 crc kubenswrapper[4646]: I1203 11:22:28.190424 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-d95tn"] Dec 03 11:22:28 crc kubenswrapper[4646]: I1203 11:22:28.195652 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-w557d"] Dec 03 11:22:28 crc kubenswrapper[4646]: I1203 11:22:28.204928 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-d2vp8"] Dec 03 11:22:28 crc kubenswrapper[4646]: I1203 11:22:28.214817 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-e410-account-create-update-2nflx"] Dec 03 11:22:29 crc kubenswrapper[4646]: I1203 11:22:29.861619 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16b24e6b-5a24-4abd-b3de-7932ce9e63c3" path="/var/lib/kubelet/pods/16b24e6b-5a24-4abd-b3de-7932ce9e63c3/volumes" Dec 03 11:22:29 crc kubenswrapper[4646]: I1203 11:22:29.862230 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76e53db6-897c-4c62-b732-034a119ab559" path="/var/lib/kubelet/pods/76e53db6-897c-4c62-b732-034a119ab559/volumes" Dec 03 11:22:29 crc kubenswrapper[4646]: I1203 11:22:29.862747 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b26f36df-eade-45a7-84e5-a181b4b3d43f" path="/var/lib/kubelet/pods/b26f36df-eade-45a7-84e5-a181b4b3d43f/volumes" Dec 03 11:22:29 crc kubenswrapper[4646]: I1203 11:22:29.863264 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5b8e5a9-ea00-4d58-8e49-50c72b9679de" path="/var/lib/kubelet/pods/b5b8e5a9-ea00-4d58-8e49-50c72b9679de/volumes" Dec 03 11:22:29 crc kubenswrapper[4646]: I1203 11:22:29.864641 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cec4fe45-3c8d-4dea-b5af-7fe6cff4594d" path="/var/lib/kubelet/pods/cec4fe45-3c8d-4dea-b5af-7fe6cff4594d/volumes" Dec 03 11:22:29 crc kubenswrapper[4646]: I1203 11:22:29.865198 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc578e46-8ab2-4567-91ba-66462a04a328" path="/var/lib/kubelet/pods/dc578e46-8ab2-4567-91ba-66462a04a328/volumes" Dec 03 11:22:32 crc kubenswrapper[4646]: I1203 11:22:32.028374 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-jpkbq"] Dec 03 11:22:32 crc kubenswrapper[4646]: I1203 11:22:32.038291 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-jpkbq"] Dec 03 11:22:33 crc kubenswrapper[4646]: I1203 11:22:33.031615 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-tjvwv"] Dec 03 11:22:33 crc kubenswrapper[4646]: I1203 11:22:33.040045 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-tjvwv"] Dec 03 11:22:33 crc kubenswrapper[4646]: I1203 11:22:33.859508 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1785e265-f0f0-4e8f-8185-8f84b3121f3b" path="/var/lib/kubelet/pods/1785e265-f0f0-4e8f-8185-8f84b3121f3b/volumes" Dec 03 11:22:33 crc kubenswrapper[4646]: I1203 11:22:33.860733 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f23e454b-e13d-47f0-a3ba-2e96f3cc0649" path="/var/lib/kubelet/pods/f23e454b-e13d-47f0-a3ba-2e96f3cc0649/volumes" Dec 03 11:22:34 crc kubenswrapper[4646]: I1203 11:22:34.848953 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:22:34 crc kubenswrapper[4646]: E1203 11:22:34.849220 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:22:49 crc kubenswrapper[4646]: I1203 11:22:49.848600 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:22:49 crc kubenswrapper[4646]: E1203 11:22:49.849234 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:23:00 crc kubenswrapper[4646]: I1203 11:23:00.848532 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:23:00 crc kubenswrapper[4646]: E1203 11:23:00.849183 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:23:03 crc kubenswrapper[4646]: I1203 11:23:03.911734 4646 scope.go:117] "RemoveContainer" containerID="f3fcb07e0073c0900914442b0fd7bbbdfd572f596ff710351ab0ff11647202cb" Dec 03 11:23:03 crc kubenswrapper[4646]: I1203 11:23:03.935828 4646 scope.go:117] "RemoveContainer" containerID="0d7671ac9b7d09df4b158f59a928970acc197aae0d07f47ff0706f352fc827a6" Dec 03 11:23:03 crc kubenswrapper[4646]: I1203 11:23:03.983040 4646 scope.go:117] "RemoveContainer" containerID="6d38eb32c115ee6ec2bbacc5d9d20a10414d18e0d52ce94ec15a3e4909f22f30" Dec 03 11:23:04 crc kubenswrapper[4646]: I1203 11:23:04.020199 4646 scope.go:117] "RemoveContainer" containerID="eb2590b56fd22c847e656d258a6487c8002c1fbefbf53a59191a36239cd17cbc" Dec 03 11:23:04 crc kubenswrapper[4646]: I1203 11:23:04.061440 4646 scope.go:117] "RemoveContainer" containerID="f82c1610661677ce6cf8fa358dde61946710745257fd8990c4c0b9a5b1b869b5" Dec 03 11:23:04 crc kubenswrapper[4646]: I1203 11:23:04.116674 4646 scope.go:117] "RemoveContainer" containerID="7f0367103f2eebf9a3973bccc99439fc2683f64f1856bb2eaac28b616dcdab13" Dec 03 11:23:04 crc kubenswrapper[4646]: I1203 11:23:04.163914 4646 scope.go:117] "RemoveContainer" containerID="8e45b35ae564b0b4890baacb3dd5dc19c5c7f59a2a9e10905e30881d50b17839" Dec 03 11:23:04 crc kubenswrapper[4646]: I1203 11:23:04.198957 4646 scope.go:117] "RemoveContainer" containerID="20fcc2042ce42d7c3777f9b9634cfe0591710b4c0454a454d153e494e89fcb3d" Dec 03 11:23:04 crc kubenswrapper[4646]: I1203 11:23:04.227402 4646 scope.go:117] "RemoveContainer" containerID="bef107b3ef6f74e23fa5762c781ac1d36aaf0e7e4048eaa6ba8c0e0e64b9c3c7" Dec 03 11:23:04 crc kubenswrapper[4646]: I1203 11:23:04.249729 4646 scope.go:117] "RemoveContainer" containerID="9bf0500a838c95634e80cfd84593d031d95f1908712bfc871fb8e341773799f7" Dec 03 11:23:04 crc kubenswrapper[4646]: I1203 11:23:04.282239 4646 scope.go:117] "RemoveContainer" containerID="d606f21242655a18dbee2c322252b15d732ae0b5a80c672b4ea931a8386148ae" Dec 03 11:23:04 crc kubenswrapper[4646]: I1203 11:23:04.314627 4646 scope.go:117] "RemoveContainer" containerID="4a006684f642fe4d5cc6d6b343aff2dfcf71766df201a33982a877ed08d34ecd" Dec 03 11:23:04 crc kubenswrapper[4646]: I1203 11:23:04.345591 4646 scope.go:117] "RemoveContainer" containerID="fbdb506e1582b61bbcd4f900d98a68fb89f18c9a3ac975b3cbe49dfec75563e2" Dec 03 11:23:04 crc kubenswrapper[4646]: I1203 11:23:04.379273 4646 scope.go:117] "RemoveContainer" containerID="560e3e176f09793233b30e3b809ad7455b3c2a1e75a4e698baf8e459815a8207" Dec 03 11:23:07 crc kubenswrapper[4646]: I1203 11:23:07.049119 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-ghphj"] Dec 03 11:23:07 crc kubenswrapper[4646]: I1203 11:23:07.059725 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-92v4c"] Dec 03 11:23:07 crc kubenswrapper[4646]: I1203 11:23:07.074062 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-ghphj"] Dec 03 11:23:07 crc kubenswrapper[4646]: I1203 11:23:07.085397 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-92v4c"] Dec 03 11:23:07 crc kubenswrapper[4646]: I1203 11:23:07.860611 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36606312-32d1-4c3a-be99-5477047af153" path="/var/lib/kubelet/pods/36606312-32d1-4c3a-be99-5477047af153/volumes" Dec 03 11:23:07 crc kubenswrapper[4646]: I1203 11:23:07.861488 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0028dab-487a-4f33-be6b-0f082d4f617e" path="/var/lib/kubelet/pods/c0028dab-487a-4f33-be6b-0f082d4f617e/volumes" Dec 03 11:23:08 crc kubenswrapper[4646]: I1203 11:23:08.031171 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-2nh9f"] Dec 03 11:23:08 crc kubenswrapper[4646]: I1203 11:23:08.038549 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-2nh9f"] Dec 03 11:23:09 crc kubenswrapper[4646]: I1203 11:23:09.859715 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aad954e4-5d92-4b2b-9771-a5f78f11d169" path="/var/lib/kubelet/pods/aad954e4-5d92-4b2b-9771-a5f78f11d169/volumes" Dec 03 11:23:11 crc kubenswrapper[4646]: I1203 11:23:11.848831 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:23:11 crc kubenswrapper[4646]: E1203 11:23:11.849601 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:23:20 crc kubenswrapper[4646]: I1203 11:23:20.046069 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-h4hq8"] Dec 03 11:23:20 crc kubenswrapper[4646]: I1203 11:23:20.053959 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-h4hq8"] Dec 03 11:23:21 crc kubenswrapper[4646]: I1203 11:23:21.873298 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="411d5bc2-b69a-4b93-bb7e-c8990b44e79a" path="/var/lib/kubelet/pods/411d5bc2-b69a-4b93-bb7e-c8990b44e79a/volumes" Dec 03 11:23:22 crc kubenswrapper[4646]: I1203 11:23:22.848094 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:23:22 crc kubenswrapper[4646]: E1203 11:23:22.849086 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:23:29 crc kubenswrapper[4646]: I1203 11:23:29.068383 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-rtwcp"] Dec 03 11:23:29 crc kubenswrapper[4646]: I1203 11:23:29.081569 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-rtwcp"] Dec 03 11:23:29 crc kubenswrapper[4646]: I1203 11:23:29.860738 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76443d6c-7930-4c62-abbb-d7ff831232ca" path="/var/lib/kubelet/pods/76443d6c-7930-4c62-abbb-d7ff831232ca/volumes" Dec 03 11:23:33 crc kubenswrapper[4646]: I1203 11:23:33.441914 4646 generic.go:334] "Generic (PLEG): container finished" podID="6185a07d-c02f-42b5-b060-c5317850e67b" containerID="9845dcf8a51e9171a44b882eb3e2c6081706b8c7481958c4bcddb26045b28539" exitCode=0 Dec 03 11:23:33 crc kubenswrapper[4646]: I1203 11:23:33.441950 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" event={"ID":"6185a07d-c02f-42b5-b060-c5317850e67b","Type":"ContainerDied","Data":"9845dcf8a51e9171a44b882eb3e2c6081706b8c7481958c4bcddb26045b28539"} Dec 03 11:23:34 crc kubenswrapper[4646]: I1203 11:23:34.840364 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" Dec 03 11:23:34 crc kubenswrapper[4646]: I1203 11:23:34.849409 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:23:34 crc kubenswrapper[4646]: E1203 11:23:34.849875 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:23:34 crc kubenswrapper[4646]: I1203 11:23:34.966653 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msqgs\" (UniqueName: \"kubernetes.io/projected/6185a07d-c02f-42b5-b060-c5317850e67b-kube-api-access-msqgs\") pod \"6185a07d-c02f-42b5-b060-c5317850e67b\" (UID: \"6185a07d-c02f-42b5-b060-c5317850e67b\") " Dec 03 11:23:34 crc kubenswrapper[4646]: I1203 11:23:34.966765 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6185a07d-c02f-42b5-b060-c5317850e67b-inventory\") pod \"6185a07d-c02f-42b5-b060-c5317850e67b\" (UID: \"6185a07d-c02f-42b5-b060-c5317850e67b\") " Dec 03 11:23:34 crc kubenswrapper[4646]: I1203 11:23:34.966795 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6185a07d-c02f-42b5-b060-c5317850e67b-ssh-key\") pod \"6185a07d-c02f-42b5-b060-c5317850e67b\" (UID: \"6185a07d-c02f-42b5-b060-c5317850e67b\") " Dec 03 11:23:34 crc kubenswrapper[4646]: I1203 11:23:34.974163 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6185a07d-c02f-42b5-b060-c5317850e67b-kube-api-access-msqgs" (OuterVolumeSpecName: "kube-api-access-msqgs") pod "6185a07d-c02f-42b5-b060-c5317850e67b" (UID: "6185a07d-c02f-42b5-b060-c5317850e67b"). InnerVolumeSpecName "kube-api-access-msqgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:23:34 crc kubenswrapper[4646]: I1203 11:23:34.996185 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6185a07d-c02f-42b5-b060-c5317850e67b-inventory" (OuterVolumeSpecName: "inventory") pod "6185a07d-c02f-42b5-b060-c5317850e67b" (UID: "6185a07d-c02f-42b5-b060-c5317850e67b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.005232 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6185a07d-c02f-42b5-b060-c5317850e67b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6185a07d-c02f-42b5-b060-c5317850e67b" (UID: "6185a07d-c02f-42b5-b060-c5317850e67b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.068664 4646 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6185a07d-c02f-42b5-b060-c5317850e67b-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.068698 4646 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6185a07d-c02f-42b5-b060-c5317850e67b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.068709 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msqgs\" (UniqueName: \"kubernetes.io/projected/6185a07d-c02f-42b5-b060-c5317850e67b-kube-api-access-msqgs\") on node \"crc\" DevicePath \"\"" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.460060 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" event={"ID":"6185a07d-c02f-42b5-b060-c5317850e67b","Type":"ContainerDied","Data":"5844d798323842ee5b6f6b2a53b8d2b5cd33c231dd845780b3e3cd6615c1e51e"} Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.460117 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.460133 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5844d798323842ee5b6f6b2a53b8d2b5cd33c231dd845780b3e3cd6615c1e51e" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.555761 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n"] Dec 03 11:23:35 crc kubenswrapper[4646]: E1203 11:23:35.556217 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6185a07d-c02f-42b5-b060-c5317850e67b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.556238 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="6185a07d-c02f-42b5-b060-c5317850e67b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.556474 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="6185a07d-c02f-42b5-b060-c5317850e67b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.557229 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.564065 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.564152 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bfp25" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.564379 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.564491 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.571271 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n"] Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.684981 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx2sg\" (UniqueName: \"kubernetes.io/projected/9a1511a5-3474-484d-a936-3ed46c8b3ed2-kube-api-access-bx2sg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n\" (UID: \"9a1511a5-3474-484d-a936-3ed46c8b3ed2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.685101 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a1511a5-3474-484d-a936-3ed46c8b3ed2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n\" (UID: \"9a1511a5-3474-484d-a936-3ed46c8b3ed2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.685130 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a1511a5-3474-484d-a936-3ed46c8b3ed2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n\" (UID: \"9a1511a5-3474-484d-a936-3ed46c8b3ed2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.787314 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx2sg\" (UniqueName: \"kubernetes.io/projected/9a1511a5-3474-484d-a936-3ed46c8b3ed2-kube-api-access-bx2sg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n\" (UID: \"9a1511a5-3474-484d-a936-3ed46c8b3ed2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.787509 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a1511a5-3474-484d-a936-3ed46c8b3ed2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n\" (UID: \"9a1511a5-3474-484d-a936-3ed46c8b3ed2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.787536 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a1511a5-3474-484d-a936-3ed46c8b3ed2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n\" (UID: \"9a1511a5-3474-484d-a936-3ed46c8b3ed2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.794162 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a1511a5-3474-484d-a936-3ed46c8b3ed2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n\" (UID: \"9a1511a5-3474-484d-a936-3ed46c8b3ed2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.800011 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a1511a5-3474-484d-a936-3ed46c8b3ed2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n\" (UID: \"9a1511a5-3474-484d-a936-3ed46c8b3ed2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.806282 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx2sg\" (UniqueName: \"kubernetes.io/projected/9a1511a5-3474-484d-a936-3ed46c8b3ed2-kube-api-access-bx2sg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n\" (UID: \"9a1511a5-3474-484d-a936-3ed46c8b3ed2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" Dec 03 11:23:35 crc kubenswrapper[4646]: I1203 11:23:35.873763 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" Dec 03 11:23:36 crc kubenswrapper[4646]: I1203 11:23:36.459502 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n"] Dec 03 11:23:36 crc kubenswrapper[4646]: I1203 11:23:36.470285 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" event={"ID":"9a1511a5-3474-484d-a936-3ed46c8b3ed2","Type":"ContainerStarted","Data":"944ea1dd0ef185000b9cbd03ee342989f50da3c1f3614573e34706995b9e1102"} Dec 03 11:23:37 crc kubenswrapper[4646]: I1203 11:23:37.481790 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" event={"ID":"9a1511a5-3474-484d-a936-3ed46c8b3ed2","Type":"ContainerStarted","Data":"1cf50de35e47dd11d8c1613217e00c758d8c32e5e0de03c9ec7ef183f430ef38"} Dec 03 11:23:37 crc kubenswrapper[4646]: I1203 11:23:37.504875 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" podStartSLOduration=2.010121471 podStartE2EDuration="2.504859266s" podCreationTimestamp="2025-12-03 11:23:35 +0000 UTC" firstStartedPulling="2025-12-03 11:23:36.463833827 +0000 UTC m=+1792.926889962" lastFinishedPulling="2025-12-03 11:23:36.958571622 +0000 UTC m=+1793.421627757" observedRunningTime="2025-12-03 11:23:37.501684904 +0000 UTC m=+1793.964741049" watchObservedRunningTime="2025-12-03 11:23:37.504859266 +0000 UTC m=+1793.967915401" Dec 03 11:23:43 crc kubenswrapper[4646]: I1203 11:23:43.528673 4646 generic.go:334] "Generic (PLEG): container finished" podID="9a1511a5-3474-484d-a936-3ed46c8b3ed2" containerID="1cf50de35e47dd11d8c1613217e00c758d8c32e5e0de03c9ec7ef183f430ef38" exitCode=0 Dec 03 11:23:43 crc kubenswrapper[4646]: I1203 11:23:43.528746 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" event={"ID":"9a1511a5-3474-484d-a936-3ed46c8b3ed2","Type":"ContainerDied","Data":"1cf50de35e47dd11d8c1613217e00c758d8c32e5e0de03c9ec7ef183f430ef38"} Dec 03 11:23:44 crc kubenswrapper[4646]: I1203 11:23:44.918828 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" Dec 03 11:23:44 crc kubenswrapper[4646]: I1203 11:23:44.971454 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a1511a5-3474-484d-a936-3ed46c8b3ed2-inventory\") pod \"9a1511a5-3474-484d-a936-3ed46c8b3ed2\" (UID: \"9a1511a5-3474-484d-a936-3ed46c8b3ed2\") " Dec 03 11:23:44 crc kubenswrapper[4646]: I1203 11:23:44.971500 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bx2sg\" (UniqueName: \"kubernetes.io/projected/9a1511a5-3474-484d-a936-3ed46c8b3ed2-kube-api-access-bx2sg\") pod \"9a1511a5-3474-484d-a936-3ed46c8b3ed2\" (UID: \"9a1511a5-3474-484d-a936-3ed46c8b3ed2\") " Dec 03 11:23:44 crc kubenswrapper[4646]: I1203 11:23:44.971656 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a1511a5-3474-484d-a936-3ed46c8b3ed2-ssh-key\") pod \"9a1511a5-3474-484d-a936-3ed46c8b3ed2\" (UID: \"9a1511a5-3474-484d-a936-3ed46c8b3ed2\") " Dec 03 11:23:44 crc kubenswrapper[4646]: I1203 11:23:44.977176 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a1511a5-3474-484d-a936-3ed46c8b3ed2-kube-api-access-bx2sg" (OuterVolumeSpecName: "kube-api-access-bx2sg") pod "9a1511a5-3474-484d-a936-3ed46c8b3ed2" (UID: "9a1511a5-3474-484d-a936-3ed46c8b3ed2"). InnerVolumeSpecName "kube-api-access-bx2sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.003004 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a1511a5-3474-484d-a936-3ed46c8b3ed2-inventory" (OuterVolumeSpecName: "inventory") pod "9a1511a5-3474-484d-a936-3ed46c8b3ed2" (UID: "9a1511a5-3474-484d-a936-3ed46c8b3ed2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.009508 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a1511a5-3474-484d-a936-3ed46c8b3ed2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9a1511a5-3474-484d-a936-3ed46c8b3ed2" (UID: "9a1511a5-3474-484d-a936-3ed46c8b3ed2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.074191 4646 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a1511a5-3474-484d-a936-3ed46c8b3ed2-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.074233 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bx2sg\" (UniqueName: \"kubernetes.io/projected/9a1511a5-3474-484d-a936-3ed46c8b3ed2-kube-api-access-bx2sg\") on node \"crc\" DevicePath \"\"" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.074251 4646 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a1511a5-3474-484d-a936-3ed46c8b3ed2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.547499 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" event={"ID":"9a1511a5-3474-484d-a936-3ed46c8b3ed2","Type":"ContainerDied","Data":"944ea1dd0ef185000b9cbd03ee342989f50da3c1f3614573e34706995b9e1102"} Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.548022 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="944ea1dd0ef185000b9cbd03ee342989f50da3c1f3614573e34706995b9e1102" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.547744 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.615198 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6"] Dec 03 11:23:45 crc kubenswrapper[4646]: E1203 11:23:45.615585 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a1511a5-3474-484d-a936-3ed46c8b3ed2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.615604 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a1511a5-3474-484d-a936-3ed46c8b3ed2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.615801 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a1511a5-3474-484d-a936-3ed46c8b3ed2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.617204 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.627320 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.627622 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.627691 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bfp25" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.627887 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.653851 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6"] Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.685734 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hq8f\" (UniqueName: \"kubernetes.io/projected/66d26829-4e09-477d-920a-07908ee19a0a-kube-api-access-6hq8f\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dn9j6\" (UID: \"66d26829-4e09-477d-920a-07908ee19a0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.685906 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66d26829-4e09-477d-920a-07908ee19a0a-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dn9j6\" (UID: \"66d26829-4e09-477d-920a-07908ee19a0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.685942 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66d26829-4e09-477d-920a-07908ee19a0a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dn9j6\" (UID: \"66d26829-4e09-477d-920a-07908ee19a0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.787642 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66d26829-4e09-477d-920a-07908ee19a0a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dn9j6\" (UID: \"66d26829-4e09-477d-920a-07908ee19a0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.787761 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hq8f\" (UniqueName: \"kubernetes.io/projected/66d26829-4e09-477d-920a-07908ee19a0a-kube-api-access-6hq8f\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dn9j6\" (UID: \"66d26829-4e09-477d-920a-07908ee19a0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.788193 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66d26829-4e09-477d-920a-07908ee19a0a-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dn9j6\" (UID: \"66d26829-4e09-477d-920a-07908ee19a0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.792854 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66d26829-4e09-477d-920a-07908ee19a0a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dn9j6\" (UID: \"66d26829-4e09-477d-920a-07908ee19a0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.796881 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66d26829-4e09-477d-920a-07908ee19a0a-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dn9j6\" (UID: \"66d26829-4e09-477d-920a-07908ee19a0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.814548 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hq8f\" (UniqueName: \"kubernetes.io/projected/66d26829-4e09-477d-920a-07908ee19a0a-kube-api-access-6hq8f\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dn9j6\" (UID: \"66d26829-4e09-477d-920a-07908ee19a0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" Dec 03 11:23:45 crc kubenswrapper[4646]: I1203 11:23:45.944240 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" Dec 03 11:23:46 crc kubenswrapper[4646]: I1203 11:23:46.513258 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6"] Dec 03 11:23:46 crc kubenswrapper[4646]: I1203 11:23:46.563728 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" event={"ID":"66d26829-4e09-477d-920a-07908ee19a0a","Type":"ContainerStarted","Data":"e022ece360f97e5789fd32d7dccde9f17abf66fd94db1fe242c57f59eaef654e"} Dec 03 11:23:47 crc kubenswrapper[4646]: I1203 11:23:47.574786 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" event={"ID":"66d26829-4e09-477d-920a-07908ee19a0a","Type":"ContainerStarted","Data":"61c69b28f6d20f7bd2c2c03d60a26527b4d1a4a6c4e8dc3828051a77d1c197fd"} Dec 03 11:23:47 crc kubenswrapper[4646]: I1203 11:23:47.601962 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" podStartSLOduration=2.000200726 podStartE2EDuration="2.601945885s" podCreationTimestamp="2025-12-03 11:23:45 +0000 UTC" firstStartedPulling="2025-12-03 11:23:46.521204529 +0000 UTC m=+1802.984260664" lastFinishedPulling="2025-12-03 11:23:47.122949688 +0000 UTC m=+1803.586005823" observedRunningTime="2025-12-03 11:23:47.599985008 +0000 UTC m=+1804.063041133" watchObservedRunningTime="2025-12-03 11:23:47.601945885 +0000 UTC m=+1804.065002020" Dec 03 11:23:47 crc kubenswrapper[4646]: I1203 11:23:47.849171 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:23:47 crc kubenswrapper[4646]: E1203 11:23:47.849417 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:24:00 crc kubenswrapper[4646]: I1203 11:24:00.849251 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:24:00 crc kubenswrapper[4646]: E1203 11:24:00.850100 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:24:04 crc kubenswrapper[4646]: I1203 11:24:04.633974 4646 scope.go:117] "RemoveContainer" containerID="a479fcfefd601ad29d040fd8ef9c169043f069b11742878431c036cead9d06d6" Dec 03 11:24:04 crc kubenswrapper[4646]: I1203 11:24:04.674381 4646 scope.go:117] "RemoveContainer" containerID="0b73033a13582c0cbbc6b90aa28c4838f5894f967db61e4e58282e22bebaab17" Dec 03 11:24:04 crc kubenswrapper[4646]: I1203 11:24:04.735811 4646 scope.go:117] "RemoveContainer" containerID="e7a283c6e0898451966968b2ef55a759b711684e8bdf653e930f30c52b97ed19" Dec 03 11:24:04 crc kubenswrapper[4646]: I1203 11:24:04.779179 4646 scope.go:117] "RemoveContainer" containerID="d541474f3e6ba25e699284ba5c49a45e9a46fb9220c76ed37b6440435a73e182" Dec 03 11:24:04 crc kubenswrapper[4646]: I1203 11:24:04.830596 4646 scope.go:117] "RemoveContainer" containerID="2b9c5c1bdd25667984f4878e0cb31ba5d3c44d04d285383a2d53bea7cb8c68a5" Dec 03 11:24:15 crc kubenswrapper[4646]: I1203 11:24:15.849833 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:24:15 crc kubenswrapper[4646]: E1203 11:24:15.850629 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:24:20 crc kubenswrapper[4646]: I1203 11:24:20.426903 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kgf8l"] Dec 03 11:24:20 crc kubenswrapper[4646]: I1203 11:24:20.430945 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:20 crc kubenswrapper[4646]: I1203 11:24:20.449722 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kgf8l"] Dec 03 11:24:20 crc kubenswrapper[4646]: I1203 11:24:20.633170 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb07edcd-e929-4361-8d8d-434c02edbe32-utilities\") pod \"redhat-operators-kgf8l\" (UID: \"cb07edcd-e929-4361-8d8d-434c02edbe32\") " pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:20 crc kubenswrapper[4646]: I1203 11:24:20.633262 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2jd7\" (UniqueName: \"kubernetes.io/projected/cb07edcd-e929-4361-8d8d-434c02edbe32-kube-api-access-q2jd7\") pod \"redhat-operators-kgf8l\" (UID: \"cb07edcd-e929-4361-8d8d-434c02edbe32\") " pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:20 crc kubenswrapper[4646]: I1203 11:24:20.633431 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb07edcd-e929-4361-8d8d-434c02edbe32-catalog-content\") pod \"redhat-operators-kgf8l\" (UID: \"cb07edcd-e929-4361-8d8d-434c02edbe32\") " pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:20 crc kubenswrapper[4646]: I1203 11:24:20.735072 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb07edcd-e929-4361-8d8d-434c02edbe32-utilities\") pod \"redhat-operators-kgf8l\" (UID: \"cb07edcd-e929-4361-8d8d-434c02edbe32\") " pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:20 crc kubenswrapper[4646]: I1203 11:24:20.735177 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2jd7\" (UniqueName: \"kubernetes.io/projected/cb07edcd-e929-4361-8d8d-434c02edbe32-kube-api-access-q2jd7\") pod \"redhat-operators-kgf8l\" (UID: \"cb07edcd-e929-4361-8d8d-434c02edbe32\") " pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:20 crc kubenswrapper[4646]: I1203 11:24:20.735312 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb07edcd-e929-4361-8d8d-434c02edbe32-catalog-content\") pod \"redhat-operators-kgf8l\" (UID: \"cb07edcd-e929-4361-8d8d-434c02edbe32\") " pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:20 crc kubenswrapper[4646]: I1203 11:24:20.735706 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb07edcd-e929-4361-8d8d-434c02edbe32-utilities\") pod \"redhat-operators-kgf8l\" (UID: \"cb07edcd-e929-4361-8d8d-434c02edbe32\") " pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:20 crc kubenswrapper[4646]: I1203 11:24:20.736838 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb07edcd-e929-4361-8d8d-434c02edbe32-catalog-content\") pod \"redhat-operators-kgf8l\" (UID: \"cb07edcd-e929-4361-8d8d-434c02edbe32\") " pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:20 crc kubenswrapper[4646]: I1203 11:24:20.763739 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2jd7\" (UniqueName: \"kubernetes.io/projected/cb07edcd-e929-4361-8d8d-434c02edbe32-kube-api-access-q2jd7\") pod \"redhat-operators-kgf8l\" (UID: \"cb07edcd-e929-4361-8d8d-434c02edbe32\") " pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.036289 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-jrbcg"] Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.045278 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-jrbcg"] Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.053755 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.429167 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7mlss"] Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.431131 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.447858 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcgqh\" (UniqueName: \"kubernetes.io/projected/66d3addd-a4ab-4179-ba7a-b53b3599dd65-kube-api-access-vcgqh\") pod \"certified-operators-7mlss\" (UID: \"66d3addd-a4ab-4179-ba7a-b53b3599dd65\") " pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.448016 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66d3addd-a4ab-4179-ba7a-b53b3599dd65-utilities\") pod \"certified-operators-7mlss\" (UID: \"66d3addd-a4ab-4179-ba7a-b53b3599dd65\") " pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.448087 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66d3addd-a4ab-4179-ba7a-b53b3599dd65-catalog-content\") pod \"certified-operators-7mlss\" (UID: \"66d3addd-a4ab-4179-ba7a-b53b3599dd65\") " pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.454508 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7mlss"] Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.531485 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kgf8l"] Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.549320 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66d3addd-a4ab-4179-ba7a-b53b3599dd65-utilities\") pod \"certified-operators-7mlss\" (UID: \"66d3addd-a4ab-4179-ba7a-b53b3599dd65\") " pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.549435 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66d3addd-a4ab-4179-ba7a-b53b3599dd65-catalog-content\") pod \"certified-operators-7mlss\" (UID: \"66d3addd-a4ab-4179-ba7a-b53b3599dd65\") " pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.549487 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcgqh\" (UniqueName: \"kubernetes.io/projected/66d3addd-a4ab-4179-ba7a-b53b3599dd65-kube-api-access-vcgqh\") pod \"certified-operators-7mlss\" (UID: \"66d3addd-a4ab-4179-ba7a-b53b3599dd65\") " pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.550683 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66d3addd-a4ab-4179-ba7a-b53b3599dd65-utilities\") pod \"certified-operators-7mlss\" (UID: \"66d3addd-a4ab-4179-ba7a-b53b3599dd65\") " pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.550954 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66d3addd-a4ab-4179-ba7a-b53b3599dd65-catalog-content\") pod \"certified-operators-7mlss\" (UID: \"66d3addd-a4ab-4179-ba7a-b53b3599dd65\") " pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.571005 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcgqh\" (UniqueName: \"kubernetes.io/projected/66d3addd-a4ab-4179-ba7a-b53b3599dd65-kube-api-access-vcgqh\") pod \"certified-operators-7mlss\" (UID: \"66d3addd-a4ab-4179-ba7a-b53b3599dd65\") " pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.753792 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:21 crc kubenswrapper[4646]: I1203 11:24:21.870527 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8adb3112-e79b-40a3-a906-65bd37489c7a" path="/var/lib/kubelet/pods/8adb3112-e79b-40a3-a906-65bd37489c7a/volumes" Dec 03 11:24:24 crc kubenswrapper[4646]: I1203 11:24:24.009808 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-1990-account-create-update-tc486"] Dec 03 11:24:24 crc kubenswrapper[4646]: I1203 11:24:24.019390 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-1990-account-create-update-tc486"] Dec 03 11:24:24 crc kubenswrapper[4646]: I1203 11:24:24.022465 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kgf8l" event={"ID":"cb07edcd-e929-4361-8d8d-434c02edbe32","Type":"ContainerStarted","Data":"36f464f44a00173271be19c0a15cbfaab7013c5acfe133f1d16dfbec5b437f75"} Dec 03 11:24:24 crc kubenswrapper[4646]: I1203 11:24:24.022507 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kgf8l" event={"ID":"cb07edcd-e929-4361-8d8d-434c02edbe32","Type":"ContainerStarted","Data":"62d2584a35589170210bb7fe3283024f09e392bef97f5e3e45d465472cb005a8"} Dec 03 11:24:24 crc kubenswrapper[4646]: I1203 11:24:24.032238 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-p45hd"] Dec 03 11:24:24 crc kubenswrapper[4646]: I1203 11:24:24.043450 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-4hgz5"] Dec 03 11:24:24 crc kubenswrapper[4646]: I1203 11:24:24.051536 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-7483-account-create-update-78pjg"] Dec 03 11:24:24 crc kubenswrapper[4646]: I1203 11:24:24.060931 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-563f-account-create-update-kkpjv"] Dec 03 11:24:24 crc kubenswrapper[4646]: I1203 11:24:24.104422 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-p45hd"] Dec 03 11:24:24 crc kubenswrapper[4646]: I1203 11:24:24.123531 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-563f-account-create-update-kkpjv"] Dec 03 11:24:24 crc kubenswrapper[4646]: I1203 11:24:24.131850 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-4hgz5"] Dec 03 11:24:24 crc kubenswrapper[4646]: I1203 11:24:24.139253 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-7483-account-create-update-78pjg"] Dec 03 11:24:24 crc kubenswrapper[4646]: I1203 11:24:24.392611 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7mlss"] Dec 03 11:24:25 crc kubenswrapper[4646]: I1203 11:24:25.033565 4646 generic.go:334] "Generic (PLEG): container finished" podID="66d3addd-a4ab-4179-ba7a-b53b3599dd65" containerID="53310823e003f8d25a5323ac6446a87ce98d6ba5dbafbabd6b7621d19d1c931a" exitCode=0 Dec 03 11:24:25 crc kubenswrapper[4646]: I1203 11:24:25.033640 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mlss" event={"ID":"66d3addd-a4ab-4179-ba7a-b53b3599dd65","Type":"ContainerDied","Data":"53310823e003f8d25a5323ac6446a87ce98d6ba5dbafbabd6b7621d19d1c931a"} Dec 03 11:24:25 crc kubenswrapper[4646]: I1203 11:24:25.033672 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mlss" event={"ID":"66d3addd-a4ab-4179-ba7a-b53b3599dd65","Type":"ContainerStarted","Data":"839b52ce5be7b7e2f72d30332a438c2c2f169442d25a08781eee4ad0a6ab462a"} Dec 03 11:24:25 crc kubenswrapper[4646]: I1203 11:24:25.038851 4646 generic.go:334] "Generic (PLEG): container finished" podID="cb07edcd-e929-4361-8d8d-434c02edbe32" containerID="36f464f44a00173271be19c0a15cbfaab7013c5acfe133f1d16dfbec5b437f75" exitCode=0 Dec 03 11:24:25 crc kubenswrapper[4646]: I1203 11:24:25.038890 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kgf8l" event={"ID":"cb07edcd-e929-4361-8d8d-434c02edbe32","Type":"ContainerDied","Data":"36f464f44a00173271be19c0a15cbfaab7013c5acfe133f1d16dfbec5b437f75"} Dec 03 11:24:25 crc kubenswrapper[4646]: I1203 11:24:25.863097 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21f00b80-abd7-49b1-8323-1f6d72a02455" path="/var/lib/kubelet/pods/21f00b80-abd7-49b1-8323-1f6d72a02455/volumes" Dec 03 11:24:25 crc kubenswrapper[4646]: I1203 11:24:25.864384 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fda3961-30ca-44c7-94f0-591558331b7d" path="/var/lib/kubelet/pods/2fda3961-30ca-44c7-94f0-591558331b7d/volumes" Dec 03 11:24:25 crc kubenswrapper[4646]: I1203 11:24:25.865147 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e229dd0-7f9f-4705-804d-584cbb33869c" path="/var/lib/kubelet/pods/3e229dd0-7f9f-4705-804d-584cbb33869c/volumes" Dec 03 11:24:25 crc kubenswrapper[4646]: I1203 11:24:25.865947 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="406832ee-9a27-4930-a47e-0fc8cf4b3a86" path="/var/lib/kubelet/pods/406832ee-9a27-4930-a47e-0fc8cf4b3a86/volumes" Dec 03 11:24:25 crc kubenswrapper[4646]: I1203 11:24:25.870984 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfd139ae-cd72-4d79-9fd2-ad5f199a33be" path="/var/lib/kubelet/pods/cfd139ae-cd72-4d79-9fd2-ad5f199a33be/volumes" Dec 03 11:24:26 crc kubenswrapper[4646]: I1203 11:24:26.050126 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kgf8l" event={"ID":"cb07edcd-e929-4361-8d8d-434c02edbe32","Type":"ContainerStarted","Data":"aaa546ad6c60ed050a8de91e0b39e50edd421c34082a8dd704cb0ce7035a236d"} Dec 03 11:24:26 crc kubenswrapper[4646]: I1203 11:24:26.054439 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mlss" event={"ID":"66d3addd-a4ab-4179-ba7a-b53b3599dd65","Type":"ContainerStarted","Data":"15d8855ac42f18e82e84ec09426564bed336d4c03219d1ad1839f06cfe4f5b74"} Dec 03 11:24:28 crc kubenswrapper[4646]: I1203 11:24:28.848504 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:24:28 crc kubenswrapper[4646]: E1203 11:24:28.849418 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:24:32 crc kubenswrapper[4646]: I1203 11:24:32.102770 4646 generic.go:334] "Generic (PLEG): container finished" podID="66d3addd-a4ab-4179-ba7a-b53b3599dd65" containerID="15d8855ac42f18e82e84ec09426564bed336d4c03219d1ad1839f06cfe4f5b74" exitCode=0 Dec 03 11:24:32 crc kubenswrapper[4646]: I1203 11:24:32.102866 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mlss" event={"ID":"66d3addd-a4ab-4179-ba7a-b53b3599dd65","Type":"ContainerDied","Data":"15d8855ac42f18e82e84ec09426564bed336d4c03219d1ad1839f06cfe4f5b74"} Dec 03 11:24:34 crc kubenswrapper[4646]: I1203 11:24:34.122177 4646 generic.go:334] "Generic (PLEG): container finished" podID="cb07edcd-e929-4361-8d8d-434c02edbe32" containerID="aaa546ad6c60ed050a8de91e0b39e50edd421c34082a8dd704cb0ce7035a236d" exitCode=0 Dec 03 11:24:34 crc kubenswrapper[4646]: I1203 11:24:34.122272 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kgf8l" event={"ID":"cb07edcd-e929-4361-8d8d-434c02edbe32","Type":"ContainerDied","Data":"aaa546ad6c60ed050a8de91e0b39e50edd421c34082a8dd704cb0ce7035a236d"} Dec 03 11:24:36 crc kubenswrapper[4646]: I1203 11:24:36.140125 4646 generic.go:334] "Generic (PLEG): container finished" podID="66d26829-4e09-477d-920a-07908ee19a0a" containerID="61c69b28f6d20f7bd2c2c03d60a26527b4d1a4a6c4e8dc3828051a77d1c197fd" exitCode=0 Dec 03 11:24:36 crc kubenswrapper[4646]: I1203 11:24:36.140214 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" event={"ID":"66d26829-4e09-477d-920a-07908ee19a0a","Type":"ContainerDied","Data":"61c69b28f6d20f7bd2c2c03d60a26527b4d1a4a6c4e8dc3828051a77d1c197fd"} Dec 03 11:24:37 crc kubenswrapper[4646]: I1203 11:24:37.813233 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" Dec 03 11:24:37 crc kubenswrapper[4646]: I1203 11:24:37.978433 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66d26829-4e09-477d-920a-07908ee19a0a-ssh-key\") pod \"66d26829-4e09-477d-920a-07908ee19a0a\" (UID: \"66d26829-4e09-477d-920a-07908ee19a0a\") " Dec 03 11:24:37 crc kubenswrapper[4646]: I1203 11:24:37.978677 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66d26829-4e09-477d-920a-07908ee19a0a-inventory\") pod \"66d26829-4e09-477d-920a-07908ee19a0a\" (UID: \"66d26829-4e09-477d-920a-07908ee19a0a\") " Dec 03 11:24:37 crc kubenswrapper[4646]: I1203 11:24:37.978717 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hq8f\" (UniqueName: \"kubernetes.io/projected/66d26829-4e09-477d-920a-07908ee19a0a-kube-api-access-6hq8f\") pod \"66d26829-4e09-477d-920a-07908ee19a0a\" (UID: \"66d26829-4e09-477d-920a-07908ee19a0a\") " Dec 03 11:24:37 crc kubenswrapper[4646]: I1203 11:24:37.999767 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66d26829-4e09-477d-920a-07908ee19a0a-kube-api-access-6hq8f" (OuterVolumeSpecName: "kube-api-access-6hq8f") pod "66d26829-4e09-477d-920a-07908ee19a0a" (UID: "66d26829-4e09-477d-920a-07908ee19a0a"). InnerVolumeSpecName "kube-api-access-6hq8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.012586 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d26829-4e09-477d-920a-07908ee19a0a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "66d26829-4e09-477d-920a-07908ee19a0a" (UID: "66d26829-4e09-477d-920a-07908ee19a0a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.018546 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d26829-4e09-477d-920a-07908ee19a0a-inventory" (OuterVolumeSpecName: "inventory") pod "66d26829-4e09-477d-920a-07908ee19a0a" (UID: "66d26829-4e09-477d-920a-07908ee19a0a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.082386 4646 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66d26829-4e09-477d-920a-07908ee19a0a-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.082425 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hq8f\" (UniqueName: \"kubernetes.io/projected/66d26829-4e09-477d-920a-07908ee19a0a-kube-api-access-6hq8f\") on node \"crc\" DevicePath \"\"" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.082436 4646 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66d26829-4e09-477d-920a-07908ee19a0a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.157772 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kgf8l" event={"ID":"cb07edcd-e929-4361-8d8d-434c02edbe32","Type":"ContainerStarted","Data":"42115cb28a6d69768375bf222df59846ae4e1f3a59118a69dcd43432e94c9b49"} Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.160299 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mlss" event={"ID":"66d3addd-a4ab-4179-ba7a-b53b3599dd65","Type":"ContainerStarted","Data":"10eb3fe441fb884e095ba4fe33496106833d2e87a3c2ae5bbc4476b19d901821"} Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.164598 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" event={"ID":"66d26829-4e09-477d-920a-07908ee19a0a","Type":"ContainerDied","Data":"e022ece360f97e5789fd32d7dccde9f17abf66fd94db1fe242c57f59eaef654e"} Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.164689 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e022ece360f97e5789fd32d7dccde9f17abf66fd94db1fe242c57f59eaef654e" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.164760 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dn9j6" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.189147 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kgf8l" podStartSLOduration=4.832462943 podStartE2EDuration="18.188935265s" podCreationTimestamp="2025-12-03 11:24:20 +0000 UTC" firstStartedPulling="2025-12-03 11:24:24.036593972 +0000 UTC m=+1840.499650097" lastFinishedPulling="2025-12-03 11:24:37.393066294 +0000 UTC m=+1853.856122419" observedRunningTime="2025-12-03 11:24:38.174252882 +0000 UTC m=+1854.637309027" watchObservedRunningTime="2025-12-03 11:24:38.188935265 +0000 UTC m=+1854.651991400" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.200911 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7mlss" podStartSLOduration=5.405282863 podStartE2EDuration="17.20089661s" podCreationTimestamp="2025-12-03 11:24:21 +0000 UTC" firstStartedPulling="2025-12-03 11:24:25.035043211 +0000 UTC m=+1841.498099346" lastFinishedPulling="2025-12-03 11:24:36.830656968 +0000 UTC m=+1853.293713093" observedRunningTime="2025-12-03 11:24:38.19777597 +0000 UTC m=+1854.660832105" watchObservedRunningTime="2025-12-03 11:24:38.20089661 +0000 UTC m=+1854.663952745" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.303164 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64"] Dec 03 11:24:38 crc kubenswrapper[4646]: E1203 11:24:38.303655 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d26829-4e09-477d-920a-07908ee19a0a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.303680 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d26829-4e09-477d-920a-07908ee19a0a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.303933 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d26829-4e09-477d-920a-07908ee19a0a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.308619 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.311527 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.314896 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bfp25" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.319276 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.323150 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.329115 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64"] Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.489202 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/246b95ce-fd50-4018-b4f1-329c2c58d1fc-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64\" (UID: \"246b95ce-fd50-4018-b4f1-329c2c58d1fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.489671 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/246b95ce-fd50-4018-b4f1-329c2c58d1fc-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64\" (UID: \"246b95ce-fd50-4018-b4f1-329c2c58d1fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.489737 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gzf9\" (UniqueName: \"kubernetes.io/projected/246b95ce-fd50-4018-b4f1-329c2c58d1fc-kube-api-access-7gzf9\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64\" (UID: \"246b95ce-fd50-4018-b4f1-329c2c58d1fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.591697 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gzf9\" (UniqueName: \"kubernetes.io/projected/246b95ce-fd50-4018-b4f1-329c2c58d1fc-kube-api-access-7gzf9\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64\" (UID: \"246b95ce-fd50-4018-b4f1-329c2c58d1fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.591839 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/246b95ce-fd50-4018-b4f1-329c2c58d1fc-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64\" (UID: \"246b95ce-fd50-4018-b4f1-329c2c58d1fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.591925 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/246b95ce-fd50-4018-b4f1-329c2c58d1fc-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64\" (UID: \"246b95ce-fd50-4018-b4f1-329c2c58d1fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.599419 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/246b95ce-fd50-4018-b4f1-329c2c58d1fc-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64\" (UID: \"246b95ce-fd50-4018-b4f1-329c2c58d1fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.600012 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/246b95ce-fd50-4018-b4f1-329c2c58d1fc-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64\" (UID: \"246b95ce-fd50-4018-b4f1-329c2c58d1fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.613348 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gzf9\" (UniqueName: \"kubernetes.io/projected/246b95ce-fd50-4018-b4f1-329c2c58d1fc-kube-api-access-7gzf9\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64\" (UID: \"246b95ce-fd50-4018-b4f1-329c2c58d1fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" Dec 03 11:24:38 crc kubenswrapper[4646]: I1203 11:24:38.644141 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" Dec 03 11:24:39 crc kubenswrapper[4646]: I1203 11:24:39.233292 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64"] Dec 03 11:24:39 crc kubenswrapper[4646]: W1203 11:24:39.239550 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod246b95ce_fd50_4018_b4f1_329c2c58d1fc.slice/crio-e1d6e0b0c09deffea4cab2f5260ed887bacf5926231e9f21e4d3ac812d3c6dde WatchSource:0}: Error finding container e1d6e0b0c09deffea4cab2f5260ed887bacf5926231e9f21e4d3ac812d3c6dde: Status 404 returned error can't find the container with id e1d6e0b0c09deffea4cab2f5260ed887bacf5926231e9f21e4d3ac812d3c6dde Dec 03 11:24:40 crc kubenswrapper[4646]: I1203 11:24:40.187365 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" event={"ID":"246b95ce-fd50-4018-b4f1-329c2c58d1fc","Type":"ContainerStarted","Data":"ec0fb8f2ffdf109165d02b87e0f96ca169726231d714528d80a5115d7bd82792"} Dec 03 11:24:40 crc kubenswrapper[4646]: I1203 11:24:40.187881 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" event={"ID":"246b95ce-fd50-4018-b4f1-329c2c58d1fc","Type":"ContainerStarted","Data":"e1d6e0b0c09deffea4cab2f5260ed887bacf5926231e9f21e4d3ac812d3c6dde"} Dec 03 11:24:41 crc kubenswrapper[4646]: I1203 11:24:41.054132 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:41 crc kubenswrapper[4646]: I1203 11:24:41.054464 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:41 crc kubenswrapper[4646]: I1203 11:24:41.754732 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:41 crc kubenswrapper[4646]: I1203 11:24:41.755087 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:42 crc kubenswrapper[4646]: I1203 11:24:42.126382 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kgf8l" podUID="cb07edcd-e929-4361-8d8d-434c02edbe32" containerName="registry-server" probeResult="failure" output=< Dec 03 11:24:42 crc kubenswrapper[4646]: timeout: failed to connect service ":50051" within 1s Dec 03 11:24:42 crc kubenswrapper[4646]: > Dec 03 11:24:42 crc kubenswrapper[4646]: I1203 11:24:42.807971 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-7mlss" podUID="66d3addd-a4ab-4179-ba7a-b53b3599dd65" containerName="registry-server" probeResult="failure" output=< Dec 03 11:24:42 crc kubenswrapper[4646]: timeout: failed to connect service ":50051" within 1s Dec 03 11:24:42 crc kubenswrapper[4646]: > Dec 03 11:24:42 crc kubenswrapper[4646]: I1203 11:24:42.848262 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:24:42 crc kubenswrapper[4646]: E1203 11:24:42.848590 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:24:45 crc kubenswrapper[4646]: I1203 11:24:45.230793 4646 generic.go:334] "Generic (PLEG): container finished" podID="246b95ce-fd50-4018-b4f1-329c2c58d1fc" containerID="ec0fb8f2ffdf109165d02b87e0f96ca169726231d714528d80a5115d7bd82792" exitCode=0 Dec 03 11:24:45 crc kubenswrapper[4646]: I1203 11:24:45.230844 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" event={"ID":"246b95ce-fd50-4018-b4f1-329c2c58d1fc","Type":"ContainerDied","Data":"ec0fb8f2ffdf109165d02b87e0f96ca169726231d714528d80a5115d7bd82792"} Dec 03 11:24:46 crc kubenswrapper[4646]: I1203 11:24:46.738621 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" Dec 03 11:24:46 crc kubenswrapper[4646]: I1203 11:24:46.785912 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gzf9\" (UniqueName: \"kubernetes.io/projected/246b95ce-fd50-4018-b4f1-329c2c58d1fc-kube-api-access-7gzf9\") pod \"246b95ce-fd50-4018-b4f1-329c2c58d1fc\" (UID: \"246b95ce-fd50-4018-b4f1-329c2c58d1fc\") " Dec 03 11:24:46 crc kubenswrapper[4646]: I1203 11:24:46.786144 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/246b95ce-fd50-4018-b4f1-329c2c58d1fc-ssh-key\") pod \"246b95ce-fd50-4018-b4f1-329c2c58d1fc\" (UID: \"246b95ce-fd50-4018-b4f1-329c2c58d1fc\") " Dec 03 11:24:46 crc kubenswrapper[4646]: I1203 11:24:46.786270 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/246b95ce-fd50-4018-b4f1-329c2c58d1fc-inventory\") pod \"246b95ce-fd50-4018-b4f1-329c2c58d1fc\" (UID: \"246b95ce-fd50-4018-b4f1-329c2c58d1fc\") " Dec 03 11:24:46 crc kubenswrapper[4646]: I1203 11:24:46.792238 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/246b95ce-fd50-4018-b4f1-329c2c58d1fc-kube-api-access-7gzf9" (OuterVolumeSpecName: "kube-api-access-7gzf9") pod "246b95ce-fd50-4018-b4f1-329c2c58d1fc" (UID: "246b95ce-fd50-4018-b4f1-329c2c58d1fc"). InnerVolumeSpecName "kube-api-access-7gzf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:24:46 crc kubenswrapper[4646]: I1203 11:24:46.830548 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/246b95ce-fd50-4018-b4f1-329c2c58d1fc-inventory" (OuterVolumeSpecName: "inventory") pod "246b95ce-fd50-4018-b4f1-329c2c58d1fc" (UID: "246b95ce-fd50-4018-b4f1-329c2c58d1fc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:24:46 crc kubenswrapper[4646]: I1203 11:24:46.832503 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/246b95ce-fd50-4018-b4f1-329c2c58d1fc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "246b95ce-fd50-4018-b4f1-329c2c58d1fc" (UID: "246b95ce-fd50-4018-b4f1-329c2c58d1fc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:24:46 crc kubenswrapper[4646]: I1203 11:24:46.890670 4646 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/246b95ce-fd50-4018-b4f1-329c2c58d1fc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 11:24:46 crc kubenswrapper[4646]: I1203 11:24:46.890849 4646 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/246b95ce-fd50-4018-b4f1-329c2c58d1fc-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 11:24:46 crc kubenswrapper[4646]: I1203 11:24:46.890886 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gzf9\" (UniqueName: \"kubernetes.io/projected/246b95ce-fd50-4018-b4f1-329c2c58d1fc-kube-api-access-7gzf9\") on node \"crc\" DevicePath \"\"" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.251210 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" event={"ID":"246b95ce-fd50-4018-b4f1-329c2c58d1fc","Type":"ContainerDied","Data":"e1d6e0b0c09deffea4cab2f5260ed887bacf5926231e9f21e4d3ac812d3c6dde"} Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.251256 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1d6e0b0c09deffea4cab2f5260ed887bacf5926231e9f21e4d3ac812d3c6dde" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.251319 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.359142 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq"] Dec 03 11:24:47 crc kubenswrapper[4646]: E1203 11:24:47.359593 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="246b95ce-fd50-4018-b4f1-329c2c58d1fc" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.359617 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="246b95ce-fd50-4018-b4f1-329c2c58d1fc" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.359814 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="246b95ce-fd50-4018-b4f1-329c2c58d1fc" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.360626 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.365575 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.365842 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.368174 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bfp25" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.368835 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.370381 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq"] Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.399760 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9320c819-432d-4440-8897-659b2fa98a20-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq\" (UID: \"9320c819-432d-4440-8897-659b2fa98a20\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.399858 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9320c819-432d-4440-8897-659b2fa98a20-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq\" (UID: \"9320c819-432d-4440-8897-659b2fa98a20\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.399929 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpv2m\" (UniqueName: \"kubernetes.io/projected/9320c819-432d-4440-8897-659b2fa98a20-kube-api-access-wpv2m\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq\" (UID: \"9320c819-432d-4440-8897-659b2fa98a20\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.501404 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9320c819-432d-4440-8897-659b2fa98a20-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq\" (UID: \"9320c819-432d-4440-8897-659b2fa98a20\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.501459 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpv2m\" (UniqueName: \"kubernetes.io/projected/9320c819-432d-4440-8897-659b2fa98a20-kube-api-access-wpv2m\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq\" (UID: \"9320c819-432d-4440-8897-659b2fa98a20\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.501571 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9320c819-432d-4440-8897-659b2fa98a20-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq\" (UID: \"9320c819-432d-4440-8897-659b2fa98a20\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.505830 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9320c819-432d-4440-8897-659b2fa98a20-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq\" (UID: \"9320c819-432d-4440-8897-659b2fa98a20\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.508716 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9320c819-432d-4440-8897-659b2fa98a20-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq\" (UID: \"9320c819-432d-4440-8897-659b2fa98a20\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.527763 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpv2m\" (UniqueName: \"kubernetes.io/projected/9320c819-432d-4440-8897-659b2fa98a20-kube-api-access-wpv2m\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq\" (UID: \"9320c819-432d-4440-8897-659b2fa98a20\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" Dec 03 11:24:47 crc kubenswrapper[4646]: I1203 11:24:47.676290 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" Dec 03 11:24:48 crc kubenswrapper[4646]: I1203 11:24:48.223800 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq"] Dec 03 11:24:48 crc kubenswrapper[4646]: I1203 11:24:48.263915 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" event={"ID":"9320c819-432d-4440-8897-659b2fa98a20","Type":"ContainerStarted","Data":"8de6c973d157d1de8cf503b78b91792f5a8ab94a5a99bea2570e698604540155"} Dec 03 11:24:50 crc kubenswrapper[4646]: I1203 11:24:50.051454 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wnq98"] Dec 03 11:24:50 crc kubenswrapper[4646]: I1203 11:24:50.057523 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wnq98"] Dec 03 11:24:50 crc kubenswrapper[4646]: I1203 11:24:50.280629 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" event={"ID":"9320c819-432d-4440-8897-659b2fa98a20","Type":"ContainerStarted","Data":"e6b350747d4d3ce7ec9a661b754511e1e42034d8ca7e185b7b728f0a9d3a3828"} Dec 03 11:24:51 crc kubenswrapper[4646]: I1203 11:24:51.138414 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:51 crc kubenswrapper[4646]: I1203 11:24:51.164969 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" podStartSLOduration=3.392926132 podStartE2EDuration="4.164946166s" podCreationTimestamp="2025-12-03 11:24:47 +0000 UTC" firstStartedPulling="2025-12-03 11:24:48.239670527 +0000 UTC m=+1864.702726662" lastFinishedPulling="2025-12-03 11:24:49.011690561 +0000 UTC m=+1865.474746696" observedRunningTime="2025-12-03 11:24:50.333932941 +0000 UTC m=+1866.796989086" watchObservedRunningTime="2025-12-03 11:24:51.164946166 +0000 UTC m=+1867.628002301" Dec 03 11:24:51 crc kubenswrapper[4646]: I1203 11:24:51.193966 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:51 crc kubenswrapper[4646]: I1203 11:24:51.799521 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:51 crc kubenswrapper[4646]: I1203 11:24:51.841733 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kgf8l"] Dec 03 11:24:51 crc kubenswrapper[4646]: I1203 11:24:51.861764 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5eb852c6-390b-4062-8350-3d614c6e1945" path="/var/lib/kubelet/pods/5eb852c6-390b-4062-8350-3d614c6e1945/volumes" Dec 03 11:24:51 crc kubenswrapper[4646]: I1203 11:24:51.865656 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:52 crc kubenswrapper[4646]: I1203 11:24:52.295104 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kgf8l" podUID="cb07edcd-e929-4361-8d8d-434c02edbe32" containerName="registry-server" containerID="cri-o://42115cb28a6d69768375bf222df59846ae4e1f3a59118a69dcd43432e94c9b49" gracePeriod=2 Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:52.847039 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:52.932484 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2jd7\" (UniqueName: \"kubernetes.io/projected/cb07edcd-e929-4361-8d8d-434c02edbe32-kube-api-access-q2jd7\") pod \"cb07edcd-e929-4361-8d8d-434c02edbe32\" (UID: \"cb07edcd-e929-4361-8d8d-434c02edbe32\") " Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:52.932550 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb07edcd-e929-4361-8d8d-434c02edbe32-catalog-content\") pod \"cb07edcd-e929-4361-8d8d-434c02edbe32\" (UID: \"cb07edcd-e929-4361-8d8d-434c02edbe32\") " Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:52.932625 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb07edcd-e929-4361-8d8d-434c02edbe32-utilities\") pod \"cb07edcd-e929-4361-8d8d-434c02edbe32\" (UID: \"cb07edcd-e929-4361-8d8d-434c02edbe32\") " Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:52.933650 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb07edcd-e929-4361-8d8d-434c02edbe32-utilities" (OuterVolumeSpecName: "utilities") pod "cb07edcd-e929-4361-8d8d-434c02edbe32" (UID: "cb07edcd-e929-4361-8d8d-434c02edbe32"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:52.947695 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb07edcd-e929-4361-8d8d-434c02edbe32-kube-api-access-q2jd7" (OuterVolumeSpecName: "kube-api-access-q2jd7") pod "cb07edcd-e929-4361-8d8d-434c02edbe32" (UID: "cb07edcd-e929-4361-8d8d-434c02edbe32"). InnerVolumeSpecName "kube-api-access-q2jd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.036735 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2jd7\" (UniqueName: \"kubernetes.io/projected/cb07edcd-e929-4361-8d8d-434c02edbe32-kube-api-access-q2jd7\") on node \"crc\" DevicePath \"\"" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.036785 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb07edcd-e929-4361-8d8d-434c02edbe32-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.060055 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb07edcd-e929-4361-8d8d-434c02edbe32-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb07edcd-e929-4361-8d8d-434c02edbe32" (UID: "cb07edcd-e929-4361-8d8d-434c02edbe32"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.138556 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb07edcd-e929-4361-8d8d-434c02edbe32-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.308302 4646 generic.go:334] "Generic (PLEG): container finished" podID="cb07edcd-e929-4361-8d8d-434c02edbe32" containerID="42115cb28a6d69768375bf222df59846ae4e1f3a59118a69dcd43432e94c9b49" exitCode=0 Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.308380 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kgf8l" event={"ID":"cb07edcd-e929-4361-8d8d-434c02edbe32","Type":"ContainerDied","Data":"42115cb28a6d69768375bf222df59846ae4e1f3a59118a69dcd43432e94c9b49"} Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.308438 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kgf8l" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.308446 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kgf8l" event={"ID":"cb07edcd-e929-4361-8d8d-434c02edbe32","Type":"ContainerDied","Data":"62d2584a35589170210bb7fe3283024f09e392bef97f5e3e45d465472cb005a8"} Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.308467 4646 scope.go:117] "RemoveContainer" containerID="42115cb28a6d69768375bf222df59846ae4e1f3a59118a69dcd43432e94c9b49" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.335509 4646 scope.go:117] "RemoveContainer" containerID="aaa546ad6c60ed050a8de91e0b39e50edd421c34082a8dd704cb0ce7035a236d" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.374496 4646 scope.go:117] "RemoveContainer" containerID="36f464f44a00173271be19c0a15cbfaab7013c5acfe133f1d16dfbec5b437f75" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.377703 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kgf8l"] Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.388264 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kgf8l"] Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.405974 4646 scope.go:117] "RemoveContainer" containerID="42115cb28a6d69768375bf222df59846ae4e1f3a59118a69dcd43432e94c9b49" Dec 03 11:24:53 crc kubenswrapper[4646]: E1203 11:24:53.406902 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42115cb28a6d69768375bf222df59846ae4e1f3a59118a69dcd43432e94c9b49\": container with ID starting with 42115cb28a6d69768375bf222df59846ae4e1f3a59118a69dcd43432e94c9b49 not found: ID does not exist" containerID="42115cb28a6d69768375bf222df59846ae4e1f3a59118a69dcd43432e94c9b49" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.406934 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42115cb28a6d69768375bf222df59846ae4e1f3a59118a69dcd43432e94c9b49"} err="failed to get container status \"42115cb28a6d69768375bf222df59846ae4e1f3a59118a69dcd43432e94c9b49\": rpc error: code = NotFound desc = could not find container \"42115cb28a6d69768375bf222df59846ae4e1f3a59118a69dcd43432e94c9b49\": container with ID starting with 42115cb28a6d69768375bf222df59846ae4e1f3a59118a69dcd43432e94c9b49 not found: ID does not exist" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.406959 4646 scope.go:117] "RemoveContainer" containerID="aaa546ad6c60ed050a8de91e0b39e50edd421c34082a8dd704cb0ce7035a236d" Dec 03 11:24:53 crc kubenswrapper[4646]: E1203 11:24:53.407300 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaa546ad6c60ed050a8de91e0b39e50edd421c34082a8dd704cb0ce7035a236d\": container with ID starting with aaa546ad6c60ed050a8de91e0b39e50edd421c34082a8dd704cb0ce7035a236d not found: ID does not exist" containerID="aaa546ad6c60ed050a8de91e0b39e50edd421c34082a8dd704cb0ce7035a236d" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.407319 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaa546ad6c60ed050a8de91e0b39e50edd421c34082a8dd704cb0ce7035a236d"} err="failed to get container status \"aaa546ad6c60ed050a8de91e0b39e50edd421c34082a8dd704cb0ce7035a236d\": rpc error: code = NotFound desc = could not find container \"aaa546ad6c60ed050a8de91e0b39e50edd421c34082a8dd704cb0ce7035a236d\": container with ID starting with aaa546ad6c60ed050a8de91e0b39e50edd421c34082a8dd704cb0ce7035a236d not found: ID does not exist" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.407349 4646 scope.go:117] "RemoveContainer" containerID="36f464f44a00173271be19c0a15cbfaab7013c5acfe133f1d16dfbec5b437f75" Dec 03 11:24:53 crc kubenswrapper[4646]: E1203 11:24:53.407560 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36f464f44a00173271be19c0a15cbfaab7013c5acfe133f1d16dfbec5b437f75\": container with ID starting with 36f464f44a00173271be19c0a15cbfaab7013c5acfe133f1d16dfbec5b437f75 not found: ID does not exist" containerID="36f464f44a00173271be19c0a15cbfaab7013c5acfe133f1d16dfbec5b437f75" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.407580 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36f464f44a00173271be19c0a15cbfaab7013c5acfe133f1d16dfbec5b437f75"} err="failed to get container status \"36f464f44a00173271be19c0a15cbfaab7013c5acfe133f1d16dfbec5b437f75\": rpc error: code = NotFound desc = could not find container \"36f464f44a00173271be19c0a15cbfaab7013c5acfe133f1d16dfbec5b437f75\": container with ID starting with 36f464f44a00173271be19c0a15cbfaab7013c5acfe133f1d16dfbec5b437f75 not found: ID does not exist" Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.629757 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7mlss"] Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.630037 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7mlss" podUID="66d3addd-a4ab-4179-ba7a-b53b3599dd65" containerName="registry-server" containerID="cri-o://10eb3fe441fb884e095ba4fe33496106833d2e87a3c2ae5bbc4476b19d901821" gracePeriod=2 Dec 03 11:24:53 crc kubenswrapper[4646]: I1203 11:24:53.861823 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb07edcd-e929-4361-8d8d-434c02edbe32" path="/var/lib/kubelet/pods/cb07edcd-e929-4361-8d8d-434c02edbe32/volumes" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.155706 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.303481 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcgqh\" (UniqueName: \"kubernetes.io/projected/66d3addd-a4ab-4179-ba7a-b53b3599dd65-kube-api-access-vcgqh\") pod \"66d3addd-a4ab-4179-ba7a-b53b3599dd65\" (UID: \"66d3addd-a4ab-4179-ba7a-b53b3599dd65\") " Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.303633 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66d3addd-a4ab-4179-ba7a-b53b3599dd65-utilities\") pod \"66d3addd-a4ab-4179-ba7a-b53b3599dd65\" (UID: \"66d3addd-a4ab-4179-ba7a-b53b3599dd65\") " Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.304487 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66d3addd-a4ab-4179-ba7a-b53b3599dd65-catalog-content\") pod \"66d3addd-a4ab-4179-ba7a-b53b3599dd65\" (UID: \"66d3addd-a4ab-4179-ba7a-b53b3599dd65\") " Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.304967 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66d3addd-a4ab-4179-ba7a-b53b3599dd65-utilities" (OuterVolumeSpecName: "utilities") pod "66d3addd-a4ab-4179-ba7a-b53b3599dd65" (UID: "66d3addd-a4ab-4179-ba7a-b53b3599dd65"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.305447 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66d3addd-a4ab-4179-ba7a-b53b3599dd65-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.314190 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66d3addd-a4ab-4179-ba7a-b53b3599dd65-kube-api-access-vcgqh" (OuterVolumeSpecName: "kube-api-access-vcgqh") pod "66d3addd-a4ab-4179-ba7a-b53b3599dd65" (UID: "66d3addd-a4ab-4179-ba7a-b53b3599dd65"). InnerVolumeSpecName "kube-api-access-vcgqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.320322 4646 generic.go:334] "Generic (PLEG): container finished" podID="66d3addd-a4ab-4179-ba7a-b53b3599dd65" containerID="10eb3fe441fb884e095ba4fe33496106833d2e87a3c2ae5bbc4476b19d901821" exitCode=0 Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.320460 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mlss" event={"ID":"66d3addd-a4ab-4179-ba7a-b53b3599dd65","Type":"ContainerDied","Data":"10eb3fe441fb884e095ba4fe33496106833d2e87a3c2ae5bbc4476b19d901821"} Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.320491 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mlss" event={"ID":"66d3addd-a4ab-4179-ba7a-b53b3599dd65","Type":"ContainerDied","Data":"839b52ce5be7b7e2f72d30332a438c2c2f169442d25a08781eee4ad0a6ab462a"} Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.320513 4646 scope.go:117] "RemoveContainer" containerID="10eb3fe441fb884e095ba4fe33496106833d2e87a3c2ae5bbc4476b19d901821" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.320649 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mlss" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.362902 4646 scope.go:117] "RemoveContainer" containerID="15d8855ac42f18e82e84ec09426564bed336d4c03219d1ad1839f06cfe4f5b74" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.365609 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66d3addd-a4ab-4179-ba7a-b53b3599dd65-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66d3addd-a4ab-4179-ba7a-b53b3599dd65" (UID: "66d3addd-a4ab-4179-ba7a-b53b3599dd65"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.384507 4646 scope.go:117] "RemoveContainer" containerID="53310823e003f8d25a5323ac6446a87ce98d6ba5dbafbabd6b7621d19d1c931a" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.401389 4646 scope.go:117] "RemoveContainer" containerID="10eb3fe441fb884e095ba4fe33496106833d2e87a3c2ae5bbc4476b19d901821" Dec 03 11:24:54 crc kubenswrapper[4646]: E1203 11:24:54.401855 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10eb3fe441fb884e095ba4fe33496106833d2e87a3c2ae5bbc4476b19d901821\": container with ID starting with 10eb3fe441fb884e095ba4fe33496106833d2e87a3c2ae5bbc4476b19d901821 not found: ID does not exist" containerID="10eb3fe441fb884e095ba4fe33496106833d2e87a3c2ae5bbc4476b19d901821" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.401967 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10eb3fe441fb884e095ba4fe33496106833d2e87a3c2ae5bbc4476b19d901821"} err="failed to get container status \"10eb3fe441fb884e095ba4fe33496106833d2e87a3c2ae5bbc4476b19d901821\": rpc error: code = NotFound desc = could not find container \"10eb3fe441fb884e095ba4fe33496106833d2e87a3c2ae5bbc4476b19d901821\": container with ID starting with 10eb3fe441fb884e095ba4fe33496106833d2e87a3c2ae5bbc4476b19d901821 not found: ID does not exist" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.402096 4646 scope.go:117] "RemoveContainer" containerID="15d8855ac42f18e82e84ec09426564bed336d4c03219d1ad1839f06cfe4f5b74" Dec 03 11:24:54 crc kubenswrapper[4646]: E1203 11:24:54.403542 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15d8855ac42f18e82e84ec09426564bed336d4c03219d1ad1839f06cfe4f5b74\": container with ID starting with 15d8855ac42f18e82e84ec09426564bed336d4c03219d1ad1839f06cfe4f5b74 not found: ID does not exist" containerID="15d8855ac42f18e82e84ec09426564bed336d4c03219d1ad1839f06cfe4f5b74" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.403588 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15d8855ac42f18e82e84ec09426564bed336d4c03219d1ad1839f06cfe4f5b74"} err="failed to get container status \"15d8855ac42f18e82e84ec09426564bed336d4c03219d1ad1839f06cfe4f5b74\": rpc error: code = NotFound desc = could not find container \"15d8855ac42f18e82e84ec09426564bed336d4c03219d1ad1839f06cfe4f5b74\": container with ID starting with 15d8855ac42f18e82e84ec09426564bed336d4c03219d1ad1839f06cfe4f5b74 not found: ID does not exist" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.403618 4646 scope.go:117] "RemoveContainer" containerID="53310823e003f8d25a5323ac6446a87ce98d6ba5dbafbabd6b7621d19d1c931a" Dec 03 11:24:54 crc kubenswrapper[4646]: E1203 11:24:54.403928 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53310823e003f8d25a5323ac6446a87ce98d6ba5dbafbabd6b7621d19d1c931a\": container with ID starting with 53310823e003f8d25a5323ac6446a87ce98d6ba5dbafbabd6b7621d19d1c931a not found: ID does not exist" containerID="53310823e003f8d25a5323ac6446a87ce98d6ba5dbafbabd6b7621d19d1c931a" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.403955 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53310823e003f8d25a5323ac6446a87ce98d6ba5dbafbabd6b7621d19d1c931a"} err="failed to get container status \"53310823e003f8d25a5323ac6446a87ce98d6ba5dbafbabd6b7621d19d1c931a\": rpc error: code = NotFound desc = could not find container \"53310823e003f8d25a5323ac6446a87ce98d6ba5dbafbabd6b7621d19d1c931a\": container with ID starting with 53310823e003f8d25a5323ac6446a87ce98d6ba5dbafbabd6b7621d19d1c931a not found: ID does not exist" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.407622 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcgqh\" (UniqueName: \"kubernetes.io/projected/66d3addd-a4ab-4179-ba7a-b53b3599dd65-kube-api-access-vcgqh\") on node \"crc\" DevicePath \"\"" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.407645 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66d3addd-a4ab-4179-ba7a-b53b3599dd65-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.667715 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7mlss"] Dec 03 11:24:54 crc kubenswrapper[4646]: I1203 11:24:54.686914 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7mlss"] Dec 03 11:24:55 crc kubenswrapper[4646]: I1203 11:24:55.849524 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:24:55 crc kubenswrapper[4646]: E1203 11:24:55.850479 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:24:55 crc kubenswrapper[4646]: I1203 11:24:55.860267 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66d3addd-a4ab-4179-ba7a-b53b3599dd65" path="/var/lib/kubelet/pods/66d3addd-a4ab-4179-ba7a-b53b3599dd65/volumes" Dec 03 11:25:04 crc kubenswrapper[4646]: I1203 11:25:04.950652 4646 scope.go:117] "RemoveContainer" containerID="a3d8337b5dbac49a804da837d377dd6fe9de1d2460e833c3306196485aca5749" Dec 03 11:25:05 crc kubenswrapper[4646]: I1203 11:25:05.005435 4646 scope.go:117] "RemoveContainer" containerID="c1173e9b04403cf852617ac87961202b9faed3962ab3ffe2e24b34617fd35f02" Dec 03 11:25:05 crc kubenswrapper[4646]: I1203 11:25:05.026666 4646 scope.go:117] "RemoveContainer" containerID="e8f98fa462b3c9ef944c54836bfe32794de06a55d798726ecafd6123ec2ebfd5" Dec 03 11:25:05 crc kubenswrapper[4646]: I1203 11:25:05.094724 4646 scope.go:117] "RemoveContainer" containerID="c8e2e34e23acc940dc29e98773a2b9bdd9fd6838c95a147efd2e916d112542eb" Dec 03 11:25:05 crc kubenswrapper[4646]: I1203 11:25:05.132317 4646 scope.go:117] "RemoveContainer" containerID="e54fe6028631354001b72bf4442de6745834ddc33e3e554c3265045db9824897" Dec 03 11:25:05 crc kubenswrapper[4646]: I1203 11:25:05.176036 4646 scope.go:117] "RemoveContainer" containerID="eb202eb66867be50b3daeb058d3496ac55c25d0ce20a3c3ee90edd628c599838" Dec 03 11:25:05 crc kubenswrapper[4646]: I1203 11:25:05.244360 4646 scope.go:117] "RemoveContainer" containerID="977beaf3e97f2c5ba1aae4b5f12a9422009c9ce4eda726e04fc4d06dbe3a191b" Dec 03 11:25:09 crc kubenswrapper[4646]: I1203 11:25:09.848637 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:25:09 crc kubenswrapper[4646]: E1203 11:25:09.849571 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:25:12 crc kubenswrapper[4646]: I1203 11:25:12.049905 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-vx27k"] Dec 03 11:25:12 crc kubenswrapper[4646]: I1203 11:25:12.058140 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-vx27k"] Dec 03 11:25:13 crc kubenswrapper[4646]: I1203 11:25:13.033425 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-j4gb5"] Dec 03 11:25:13 crc kubenswrapper[4646]: I1203 11:25:13.040824 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-j4gb5"] Dec 03 11:25:13 crc kubenswrapper[4646]: I1203 11:25:13.864762 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56efccd7-660f-4052-afa3-dab6a63e93a2" path="/var/lib/kubelet/pods/56efccd7-660f-4052-afa3-dab6a63e93a2/volumes" Dec 03 11:25:13 crc kubenswrapper[4646]: I1203 11:25:13.865433 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f37ef5fc-6976-46cf-87ee-6c3d8acf75a8" path="/var/lib/kubelet/pods/f37ef5fc-6976-46cf-87ee-6c3d8acf75a8/volumes" Dec 03 11:25:24 crc kubenswrapper[4646]: I1203 11:25:24.849231 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:25:24 crc kubenswrapper[4646]: E1203 11:25:24.850154 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:25:38 crc kubenswrapper[4646]: I1203 11:25:38.849703 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:25:38 crc kubenswrapper[4646]: E1203 11:25:38.850851 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:25:50 crc kubenswrapper[4646]: I1203 11:25:50.848314 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:25:50 crc kubenswrapper[4646]: E1203 11:25:50.849491 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:25:50 crc kubenswrapper[4646]: I1203 11:25:50.852437 4646 generic.go:334] "Generic (PLEG): container finished" podID="9320c819-432d-4440-8897-659b2fa98a20" containerID="e6b350747d4d3ce7ec9a661b754511e1e42034d8ca7e185b7b728f0a9d3a3828" exitCode=0 Dec 03 11:25:50 crc kubenswrapper[4646]: I1203 11:25:50.852477 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" event={"ID":"9320c819-432d-4440-8897-659b2fa98a20","Type":"ContainerDied","Data":"e6b350747d4d3ce7ec9a661b754511e1e42034d8ca7e185b7b728f0a9d3a3828"} Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.246979 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.299016 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpv2m\" (UniqueName: \"kubernetes.io/projected/9320c819-432d-4440-8897-659b2fa98a20-kube-api-access-wpv2m\") pod \"9320c819-432d-4440-8897-659b2fa98a20\" (UID: \"9320c819-432d-4440-8897-659b2fa98a20\") " Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.299094 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9320c819-432d-4440-8897-659b2fa98a20-ssh-key\") pod \"9320c819-432d-4440-8897-659b2fa98a20\" (UID: \"9320c819-432d-4440-8897-659b2fa98a20\") " Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.299283 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9320c819-432d-4440-8897-659b2fa98a20-inventory\") pod \"9320c819-432d-4440-8897-659b2fa98a20\" (UID: \"9320c819-432d-4440-8897-659b2fa98a20\") " Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.306278 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9320c819-432d-4440-8897-659b2fa98a20-kube-api-access-wpv2m" (OuterVolumeSpecName: "kube-api-access-wpv2m") pod "9320c819-432d-4440-8897-659b2fa98a20" (UID: "9320c819-432d-4440-8897-659b2fa98a20"). InnerVolumeSpecName "kube-api-access-wpv2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.334441 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9320c819-432d-4440-8897-659b2fa98a20-inventory" (OuterVolumeSpecName: "inventory") pod "9320c819-432d-4440-8897-659b2fa98a20" (UID: "9320c819-432d-4440-8897-659b2fa98a20"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.334559 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9320c819-432d-4440-8897-659b2fa98a20-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9320c819-432d-4440-8897-659b2fa98a20" (UID: "9320c819-432d-4440-8897-659b2fa98a20"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.400931 4646 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9320c819-432d-4440-8897-659b2fa98a20-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.400977 4646 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9320c819-432d-4440-8897-659b2fa98a20-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.400987 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpv2m\" (UniqueName: \"kubernetes.io/projected/9320c819-432d-4440-8897-659b2fa98a20-kube-api-access-wpv2m\") on node \"crc\" DevicePath \"\"" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.872721 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" event={"ID":"9320c819-432d-4440-8897-659b2fa98a20","Type":"ContainerDied","Data":"8de6c973d157d1de8cf503b78b91792f5a8ab94a5a99bea2570e698604540155"} Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.873036 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8de6c973d157d1de8cf503b78b91792f5a8ab94a5a99bea2570e698604540155" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.872794 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.955213 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-7vhz8"] Dec 03 11:25:52 crc kubenswrapper[4646]: E1203 11:25:52.955577 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d3addd-a4ab-4179-ba7a-b53b3599dd65" containerName="extract-utilities" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.955594 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d3addd-a4ab-4179-ba7a-b53b3599dd65" containerName="extract-utilities" Dec 03 11:25:52 crc kubenswrapper[4646]: E1203 11:25:52.955618 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb07edcd-e929-4361-8d8d-434c02edbe32" containerName="registry-server" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.955624 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb07edcd-e929-4361-8d8d-434c02edbe32" containerName="registry-server" Dec 03 11:25:52 crc kubenswrapper[4646]: E1203 11:25:52.955635 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb07edcd-e929-4361-8d8d-434c02edbe32" containerName="extract-content" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.955642 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb07edcd-e929-4361-8d8d-434c02edbe32" containerName="extract-content" Dec 03 11:25:52 crc kubenswrapper[4646]: E1203 11:25:52.955660 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d3addd-a4ab-4179-ba7a-b53b3599dd65" containerName="registry-server" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.955666 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d3addd-a4ab-4179-ba7a-b53b3599dd65" containerName="registry-server" Dec 03 11:25:52 crc kubenswrapper[4646]: E1203 11:25:52.955684 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d3addd-a4ab-4179-ba7a-b53b3599dd65" containerName="extract-content" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.955691 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d3addd-a4ab-4179-ba7a-b53b3599dd65" containerName="extract-content" Dec 03 11:25:52 crc kubenswrapper[4646]: E1203 11:25:52.955700 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9320c819-432d-4440-8897-659b2fa98a20" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.955707 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="9320c819-432d-4440-8897-659b2fa98a20" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 11:25:52 crc kubenswrapper[4646]: E1203 11:25:52.955719 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb07edcd-e929-4361-8d8d-434c02edbe32" containerName="extract-utilities" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.955725 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb07edcd-e929-4361-8d8d-434c02edbe32" containerName="extract-utilities" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.955875 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="9320c819-432d-4440-8897-659b2fa98a20" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.955890 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d3addd-a4ab-4179-ba7a-b53b3599dd65" containerName="registry-server" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.955905 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb07edcd-e929-4361-8d8d-434c02edbe32" containerName="registry-server" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.956473 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.958666 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bfp25" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.960929 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.961095 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 11:25:52 crc kubenswrapper[4646]: I1203 11:25:52.961911 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 11:25:53 crc kubenswrapper[4646]: I1203 11:25:53.012987 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e44d032-6062-45a5-b3cd-59e54618238f-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-7vhz8\" (UID: \"5e44d032-6062-45a5-b3cd-59e54618238f\") " pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" Dec 03 11:25:53 crc kubenswrapper[4646]: I1203 11:25:53.013105 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5e44d032-6062-45a5-b3cd-59e54618238f-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-7vhz8\" (UID: \"5e44d032-6062-45a5-b3cd-59e54618238f\") " pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" Dec 03 11:25:53 crc kubenswrapper[4646]: I1203 11:25:53.013216 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d58r4\" (UniqueName: \"kubernetes.io/projected/5e44d032-6062-45a5-b3cd-59e54618238f-kube-api-access-d58r4\") pod \"ssh-known-hosts-edpm-deployment-7vhz8\" (UID: \"5e44d032-6062-45a5-b3cd-59e54618238f\") " pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" Dec 03 11:25:53 crc kubenswrapper[4646]: I1203 11:25:53.044241 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-7vhz8"] Dec 03 11:25:53 crc kubenswrapper[4646]: I1203 11:25:53.114697 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e44d032-6062-45a5-b3cd-59e54618238f-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-7vhz8\" (UID: \"5e44d032-6062-45a5-b3cd-59e54618238f\") " pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" Dec 03 11:25:53 crc kubenswrapper[4646]: I1203 11:25:53.114786 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5e44d032-6062-45a5-b3cd-59e54618238f-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-7vhz8\" (UID: \"5e44d032-6062-45a5-b3cd-59e54618238f\") " pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" Dec 03 11:25:53 crc kubenswrapper[4646]: I1203 11:25:53.114860 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d58r4\" (UniqueName: \"kubernetes.io/projected/5e44d032-6062-45a5-b3cd-59e54618238f-kube-api-access-d58r4\") pod \"ssh-known-hosts-edpm-deployment-7vhz8\" (UID: \"5e44d032-6062-45a5-b3cd-59e54618238f\") " pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" Dec 03 11:25:53 crc kubenswrapper[4646]: I1203 11:25:53.119416 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e44d032-6062-45a5-b3cd-59e54618238f-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-7vhz8\" (UID: \"5e44d032-6062-45a5-b3cd-59e54618238f\") " pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" Dec 03 11:25:53 crc kubenswrapper[4646]: I1203 11:25:53.129015 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5e44d032-6062-45a5-b3cd-59e54618238f-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-7vhz8\" (UID: \"5e44d032-6062-45a5-b3cd-59e54618238f\") " pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" Dec 03 11:25:53 crc kubenswrapper[4646]: I1203 11:25:53.136430 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d58r4\" (UniqueName: \"kubernetes.io/projected/5e44d032-6062-45a5-b3cd-59e54618238f-kube-api-access-d58r4\") pod \"ssh-known-hosts-edpm-deployment-7vhz8\" (UID: \"5e44d032-6062-45a5-b3cd-59e54618238f\") " pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" Dec 03 11:25:53 crc kubenswrapper[4646]: I1203 11:25:53.271757 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" Dec 03 11:25:53 crc kubenswrapper[4646]: I1203 11:25:53.860870 4646 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 11:25:53 crc kubenswrapper[4646]: I1203 11:25:53.870898 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-7vhz8"] Dec 03 11:25:53 crc kubenswrapper[4646]: I1203 11:25:53.887013 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" event={"ID":"5e44d032-6062-45a5-b3cd-59e54618238f","Type":"ContainerStarted","Data":"2ef54f5979eae9756438a56f2a5f9e61dd8e25c1c56f7475c54fdbbe54a20f39"} Dec 03 11:25:54 crc kubenswrapper[4646]: I1203 11:25:54.042260 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-gckcg"] Dec 03 11:25:54 crc kubenswrapper[4646]: I1203 11:25:54.050617 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-gckcg"] Dec 03 11:25:55 crc kubenswrapper[4646]: I1203 11:25:55.859603 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c171132a-33a9-4df3-8bb3-1789bb031da5" path="/var/lib/kubelet/pods/c171132a-33a9-4df3-8bb3-1789bb031da5/volumes" Dec 03 11:25:55 crc kubenswrapper[4646]: I1203 11:25:55.905221 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" event={"ID":"5e44d032-6062-45a5-b3cd-59e54618238f","Type":"ContainerStarted","Data":"bf1964350c3a1358a2d9fcc71c776e8bd108ad0a68f724e6d825cdfbfe67ef43"} Dec 03 11:25:55 crc kubenswrapper[4646]: I1203 11:25:55.924210 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" podStartSLOduration=3.042266533 podStartE2EDuration="3.924186912s" podCreationTimestamp="2025-12-03 11:25:52 +0000 UTC" firstStartedPulling="2025-12-03 11:25:53.860604876 +0000 UTC m=+1930.323661011" lastFinishedPulling="2025-12-03 11:25:54.742525235 +0000 UTC m=+1931.205581390" observedRunningTime="2025-12-03 11:25:55.922695848 +0000 UTC m=+1932.385752003" watchObservedRunningTime="2025-12-03 11:25:55.924186912 +0000 UTC m=+1932.387243047" Dec 03 11:26:02 crc kubenswrapper[4646]: I1203 11:26:02.958447 4646 generic.go:334] "Generic (PLEG): container finished" podID="5e44d032-6062-45a5-b3cd-59e54618238f" containerID="bf1964350c3a1358a2d9fcc71c776e8bd108ad0a68f724e6d825cdfbfe67ef43" exitCode=0 Dec 03 11:26:02 crc kubenswrapper[4646]: I1203 11:26:02.958501 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" event={"ID":"5e44d032-6062-45a5-b3cd-59e54618238f","Type":"ContainerDied","Data":"bf1964350c3a1358a2d9fcc71c776e8bd108ad0a68f724e6d825cdfbfe67ef43"} Dec 03 11:26:04 crc kubenswrapper[4646]: I1203 11:26:04.377182 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" Dec 03 11:26:04 crc kubenswrapper[4646]: I1203 11:26:04.458962 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d58r4\" (UniqueName: \"kubernetes.io/projected/5e44d032-6062-45a5-b3cd-59e54618238f-kube-api-access-d58r4\") pod \"5e44d032-6062-45a5-b3cd-59e54618238f\" (UID: \"5e44d032-6062-45a5-b3cd-59e54618238f\") " Dec 03 11:26:04 crc kubenswrapper[4646]: I1203 11:26:04.459406 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5e44d032-6062-45a5-b3cd-59e54618238f-inventory-0\") pod \"5e44d032-6062-45a5-b3cd-59e54618238f\" (UID: \"5e44d032-6062-45a5-b3cd-59e54618238f\") " Dec 03 11:26:04 crc kubenswrapper[4646]: I1203 11:26:04.459529 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e44d032-6062-45a5-b3cd-59e54618238f-ssh-key-openstack-edpm-ipam\") pod \"5e44d032-6062-45a5-b3cd-59e54618238f\" (UID: \"5e44d032-6062-45a5-b3cd-59e54618238f\") " Dec 03 11:26:04 crc kubenswrapper[4646]: I1203 11:26:04.465487 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e44d032-6062-45a5-b3cd-59e54618238f-kube-api-access-d58r4" (OuterVolumeSpecName: "kube-api-access-d58r4") pod "5e44d032-6062-45a5-b3cd-59e54618238f" (UID: "5e44d032-6062-45a5-b3cd-59e54618238f"). InnerVolumeSpecName "kube-api-access-d58r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:26:04 crc kubenswrapper[4646]: I1203 11:26:04.483754 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e44d032-6062-45a5-b3cd-59e54618238f-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "5e44d032-6062-45a5-b3cd-59e54618238f" (UID: "5e44d032-6062-45a5-b3cd-59e54618238f"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:26:04 crc kubenswrapper[4646]: I1203 11:26:04.485808 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e44d032-6062-45a5-b3cd-59e54618238f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5e44d032-6062-45a5-b3cd-59e54618238f" (UID: "5e44d032-6062-45a5-b3cd-59e54618238f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:26:04 crc kubenswrapper[4646]: I1203 11:26:04.561234 4646 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e44d032-6062-45a5-b3cd-59e54618238f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 11:26:04 crc kubenswrapper[4646]: I1203 11:26:04.561269 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d58r4\" (UniqueName: \"kubernetes.io/projected/5e44d032-6062-45a5-b3cd-59e54618238f-kube-api-access-d58r4\") on node \"crc\" DevicePath \"\"" Dec 03 11:26:04 crc kubenswrapper[4646]: I1203 11:26:04.561283 4646 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5e44d032-6062-45a5-b3cd-59e54618238f-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 03 11:26:04 crc kubenswrapper[4646]: I1203 11:26:04.848931 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:26:04 crc kubenswrapper[4646]: E1203 11:26:04.849195 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:26:04 crc kubenswrapper[4646]: I1203 11:26:04.975499 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" event={"ID":"5e44d032-6062-45a5-b3cd-59e54618238f","Type":"ContainerDied","Data":"2ef54f5979eae9756438a56f2a5f9e61dd8e25c1c56f7475c54fdbbe54a20f39"} Dec 03 11:26:04 crc kubenswrapper[4646]: I1203 11:26:04.975536 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ef54f5979eae9756438a56f2a5f9e61dd8e25c1c56f7475c54fdbbe54a20f39" Dec 03 11:26:04 crc kubenswrapper[4646]: I1203 11:26:04.975557 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-7vhz8" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.075303 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg"] Dec 03 11:26:05 crc kubenswrapper[4646]: E1203 11:26:05.075961 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e44d032-6062-45a5-b3cd-59e54618238f" containerName="ssh-known-hosts-edpm-deployment" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.076027 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e44d032-6062-45a5-b3cd-59e54618238f" containerName="ssh-known-hosts-edpm-deployment" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.076518 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e44d032-6062-45a5-b3cd-59e54618238f" containerName="ssh-known-hosts-edpm-deployment" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.077627 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.085393 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.085605 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bfp25" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.088594 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg"] Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.089017 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.089292 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.171982 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffvqf\" (UniqueName: \"kubernetes.io/projected/f01298f9-3c81-4297-921f-0a201d0ac557-kube-api-access-ffvqf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vsdg\" (UID: \"f01298f9-3c81-4297-921f-0a201d0ac557\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.172040 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f01298f9-3c81-4297-921f-0a201d0ac557-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vsdg\" (UID: \"f01298f9-3c81-4297-921f-0a201d0ac557\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.172572 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f01298f9-3c81-4297-921f-0a201d0ac557-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vsdg\" (UID: \"f01298f9-3c81-4297-921f-0a201d0ac557\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.274587 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffvqf\" (UniqueName: \"kubernetes.io/projected/f01298f9-3c81-4297-921f-0a201d0ac557-kube-api-access-ffvqf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vsdg\" (UID: \"f01298f9-3c81-4297-921f-0a201d0ac557\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.274638 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f01298f9-3c81-4297-921f-0a201d0ac557-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vsdg\" (UID: \"f01298f9-3c81-4297-921f-0a201d0ac557\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.274665 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f01298f9-3c81-4297-921f-0a201d0ac557-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vsdg\" (UID: \"f01298f9-3c81-4297-921f-0a201d0ac557\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.287367 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f01298f9-3c81-4297-921f-0a201d0ac557-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vsdg\" (UID: \"f01298f9-3c81-4297-921f-0a201d0ac557\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.287867 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f01298f9-3c81-4297-921f-0a201d0ac557-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vsdg\" (UID: \"f01298f9-3c81-4297-921f-0a201d0ac557\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.289975 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffvqf\" (UniqueName: \"kubernetes.io/projected/f01298f9-3c81-4297-921f-0a201d0ac557-kube-api-access-ffvqf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vsdg\" (UID: \"f01298f9-3c81-4297-921f-0a201d0ac557\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.398871 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.418702 4646 scope.go:117] "RemoveContainer" containerID="55f494c40970f92c45b900e8e523f7be3167fa07ff2296242dd792044a06cf32" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.594645 4646 scope.go:117] "RemoveContainer" containerID="7d185132f8aab22aff0d4aa5add5f3a1da579b91cebeb0b4e1f10046c1b01d55" Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.647986 4646 scope.go:117] "RemoveContainer" containerID="59ab2024684ff6d3c8ff7ab17b5b5e02ba1aef3d3411b4759799dca1d9e408b4" Dec 03 11:26:05 crc kubenswrapper[4646]: W1203 11:26:05.934038 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf01298f9_3c81_4297_921f_0a201d0ac557.slice/crio-4d02d5be395cdaa205f336ffbf3574573b7dd90599fd905857b8a516aac3f418 WatchSource:0}: Error finding container 4d02d5be395cdaa205f336ffbf3574573b7dd90599fd905857b8a516aac3f418: Status 404 returned error can't find the container with id 4d02d5be395cdaa205f336ffbf3574573b7dd90599fd905857b8a516aac3f418 Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.935195 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg"] Dec 03 11:26:05 crc kubenswrapper[4646]: I1203 11:26:05.985321 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" event={"ID":"f01298f9-3c81-4297-921f-0a201d0ac557","Type":"ContainerStarted","Data":"4d02d5be395cdaa205f336ffbf3574573b7dd90599fd905857b8a516aac3f418"} Dec 03 11:26:08 crc kubenswrapper[4646]: I1203 11:26:08.007669 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" event={"ID":"f01298f9-3c81-4297-921f-0a201d0ac557","Type":"ContainerStarted","Data":"01aa4bfa607164dd06622719bde8bdb7dce4c89cde291a933c3e593cd0741f8a"} Dec 03 11:26:08 crc kubenswrapper[4646]: I1203 11:26:08.030372 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" podStartSLOduration=1.593103142 podStartE2EDuration="3.030356198s" podCreationTimestamp="2025-12-03 11:26:05 +0000 UTC" firstStartedPulling="2025-12-03 11:26:05.936593103 +0000 UTC m=+1942.399649248" lastFinishedPulling="2025-12-03 11:26:07.373846159 +0000 UTC m=+1943.836902304" observedRunningTime="2025-12-03 11:26:08.026507566 +0000 UTC m=+1944.489563701" watchObservedRunningTime="2025-12-03 11:26:08.030356198 +0000 UTC m=+1944.493412333" Dec 03 11:26:16 crc kubenswrapper[4646]: I1203 11:26:16.849034 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:26:16 crc kubenswrapper[4646]: E1203 11:26:16.850066 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:26:17 crc kubenswrapper[4646]: I1203 11:26:17.089273 4646 generic.go:334] "Generic (PLEG): container finished" podID="f01298f9-3c81-4297-921f-0a201d0ac557" containerID="01aa4bfa607164dd06622719bde8bdb7dce4c89cde291a933c3e593cd0741f8a" exitCode=0 Dec 03 11:26:17 crc kubenswrapper[4646]: I1203 11:26:17.089391 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" event={"ID":"f01298f9-3c81-4297-921f-0a201d0ac557","Type":"ContainerDied","Data":"01aa4bfa607164dd06622719bde8bdb7dce4c89cde291a933c3e593cd0741f8a"} Dec 03 11:26:18 crc kubenswrapper[4646]: I1203 11:26:18.660927 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" Dec 03 11:26:18 crc kubenswrapper[4646]: I1203 11:26:18.715316 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f01298f9-3c81-4297-921f-0a201d0ac557-inventory\") pod \"f01298f9-3c81-4297-921f-0a201d0ac557\" (UID: \"f01298f9-3c81-4297-921f-0a201d0ac557\") " Dec 03 11:26:18 crc kubenswrapper[4646]: I1203 11:26:18.715449 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f01298f9-3c81-4297-921f-0a201d0ac557-ssh-key\") pod \"f01298f9-3c81-4297-921f-0a201d0ac557\" (UID: \"f01298f9-3c81-4297-921f-0a201d0ac557\") " Dec 03 11:26:18 crc kubenswrapper[4646]: I1203 11:26:18.715537 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffvqf\" (UniqueName: \"kubernetes.io/projected/f01298f9-3c81-4297-921f-0a201d0ac557-kube-api-access-ffvqf\") pod \"f01298f9-3c81-4297-921f-0a201d0ac557\" (UID: \"f01298f9-3c81-4297-921f-0a201d0ac557\") " Dec 03 11:26:18 crc kubenswrapper[4646]: I1203 11:26:18.720791 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f01298f9-3c81-4297-921f-0a201d0ac557-kube-api-access-ffvqf" (OuterVolumeSpecName: "kube-api-access-ffvqf") pod "f01298f9-3c81-4297-921f-0a201d0ac557" (UID: "f01298f9-3c81-4297-921f-0a201d0ac557"). InnerVolumeSpecName "kube-api-access-ffvqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:26:18 crc kubenswrapper[4646]: I1203 11:26:18.747031 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f01298f9-3c81-4297-921f-0a201d0ac557-inventory" (OuterVolumeSpecName: "inventory") pod "f01298f9-3c81-4297-921f-0a201d0ac557" (UID: "f01298f9-3c81-4297-921f-0a201d0ac557"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:26:18 crc kubenswrapper[4646]: I1203 11:26:18.755554 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f01298f9-3c81-4297-921f-0a201d0ac557-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f01298f9-3c81-4297-921f-0a201d0ac557" (UID: "f01298f9-3c81-4297-921f-0a201d0ac557"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:26:18 crc kubenswrapper[4646]: I1203 11:26:18.818208 4646 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f01298f9-3c81-4297-921f-0a201d0ac557-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 11:26:18 crc kubenswrapper[4646]: I1203 11:26:18.818254 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffvqf\" (UniqueName: \"kubernetes.io/projected/f01298f9-3c81-4297-921f-0a201d0ac557-kube-api-access-ffvqf\") on node \"crc\" DevicePath \"\"" Dec 03 11:26:18 crc kubenswrapper[4646]: I1203 11:26:18.818269 4646 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f01298f9-3c81-4297-921f-0a201d0ac557-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.107489 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" event={"ID":"f01298f9-3c81-4297-921f-0a201d0ac557","Type":"ContainerDied","Data":"4d02d5be395cdaa205f336ffbf3574573b7dd90599fd905857b8a516aac3f418"} Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.107556 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d02d5be395cdaa205f336ffbf3574573b7dd90599fd905857b8a516aac3f418" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.107618 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vsdg" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.195709 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9"] Dec 03 11:26:19 crc kubenswrapper[4646]: E1203 11:26:19.196149 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f01298f9-3c81-4297-921f-0a201d0ac557" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.196171 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="f01298f9-3c81-4297-921f-0a201d0ac557" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.196398 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="f01298f9-3c81-4297-921f-0a201d0ac557" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.197102 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.201897 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.201953 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.202131 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.204954 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-bfp25" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.209654 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9"] Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.224601 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6724a5b-6c06-48df-a742-d9c0484ab526-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9\" (UID: \"a6724a5b-6c06-48df-a742-d9c0484ab526\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.224737 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6724a5b-6c06-48df-a742-d9c0484ab526-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9\" (UID: \"a6724a5b-6c06-48df-a742-d9c0484ab526\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.224760 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvblg\" (UniqueName: \"kubernetes.io/projected/a6724a5b-6c06-48df-a742-d9c0484ab526-kube-api-access-hvblg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9\" (UID: \"a6724a5b-6c06-48df-a742-d9c0484ab526\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.325302 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6724a5b-6c06-48df-a742-d9c0484ab526-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9\" (UID: \"a6724a5b-6c06-48df-a742-d9c0484ab526\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.325469 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6724a5b-6c06-48df-a742-d9c0484ab526-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9\" (UID: \"a6724a5b-6c06-48df-a742-d9c0484ab526\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.325500 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvblg\" (UniqueName: \"kubernetes.io/projected/a6724a5b-6c06-48df-a742-d9c0484ab526-kube-api-access-hvblg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9\" (UID: \"a6724a5b-6c06-48df-a742-d9c0484ab526\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.331302 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6724a5b-6c06-48df-a742-d9c0484ab526-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9\" (UID: \"a6724a5b-6c06-48df-a742-d9c0484ab526\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.337710 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6724a5b-6c06-48df-a742-d9c0484ab526-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9\" (UID: \"a6724a5b-6c06-48df-a742-d9c0484ab526\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.341908 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvblg\" (UniqueName: \"kubernetes.io/projected/a6724a5b-6c06-48df-a742-d9c0484ab526-kube-api-access-hvblg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9\" (UID: \"a6724a5b-6c06-48df-a742-d9c0484ab526\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" Dec 03 11:26:19 crc kubenswrapper[4646]: I1203 11:26:19.522978 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" Dec 03 11:26:20 crc kubenswrapper[4646]: I1203 11:26:20.034871 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9"] Dec 03 11:26:20 crc kubenswrapper[4646]: I1203 11:26:20.117017 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" event={"ID":"a6724a5b-6c06-48df-a742-d9c0484ab526","Type":"ContainerStarted","Data":"2686b4f8bf1a07474352ae6ad8e2e15869c678fe02d758e84738abc62c028b93"} Dec 03 11:26:21 crc kubenswrapper[4646]: I1203 11:26:21.126669 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" event={"ID":"a6724a5b-6c06-48df-a742-d9c0484ab526","Type":"ContainerStarted","Data":"aef742d36227651376146b316ffdc49fb2d3fbb628e8d9b39b7574eecd75aaf9"} Dec 03 11:26:21 crc kubenswrapper[4646]: I1203 11:26:21.150661 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" podStartSLOduration=1.5664365120000001 podStartE2EDuration="2.150643597s" podCreationTimestamp="2025-12-03 11:26:19 +0000 UTC" firstStartedPulling="2025-12-03 11:26:20.037170019 +0000 UTC m=+1956.500226154" lastFinishedPulling="2025-12-03 11:26:20.621377104 +0000 UTC m=+1957.084433239" observedRunningTime="2025-12-03 11:26:21.141482275 +0000 UTC m=+1957.604538400" watchObservedRunningTime="2025-12-03 11:26:21.150643597 +0000 UTC m=+1957.613699722" Dec 03 11:26:31 crc kubenswrapper[4646]: I1203 11:26:31.850245 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:26:31 crc kubenswrapper[4646]: E1203 11:26:31.851010 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:26:32 crc kubenswrapper[4646]: I1203 11:26:32.225979 4646 generic.go:334] "Generic (PLEG): container finished" podID="a6724a5b-6c06-48df-a742-d9c0484ab526" containerID="aef742d36227651376146b316ffdc49fb2d3fbb628e8d9b39b7574eecd75aaf9" exitCode=0 Dec 03 11:26:32 crc kubenswrapper[4646]: I1203 11:26:32.226022 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" event={"ID":"a6724a5b-6c06-48df-a742-d9c0484ab526","Type":"ContainerDied","Data":"aef742d36227651376146b316ffdc49fb2d3fbb628e8d9b39b7574eecd75aaf9"} Dec 03 11:26:33 crc kubenswrapper[4646]: I1203 11:26:33.646695 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" Dec 03 11:26:33 crc kubenswrapper[4646]: I1203 11:26:33.837214 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvblg\" (UniqueName: \"kubernetes.io/projected/a6724a5b-6c06-48df-a742-d9c0484ab526-kube-api-access-hvblg\") pod \"a6724a5b-6c06-48df-a742-d9c0484ab526\" (UID: \"a6724a5b-6c06-48df-a742-d9c0484ab526\") " Dec 03 11:26:33 crc kubenswrapper[4646]: I1203 11:26:33.837322 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6724a5b-6c06-48df-a742-d9c0484ab526-inventory\") pod \"a6724a5b-6c06-48df-a742-d9c0484ab526\" (UID: \"a6724a5b-6c06-48df-a742-d9c0484ab526\") " Dec 03 11:26:33 crc kubenswrapper[4646]: I1203 11:26:33.837399 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6724a5b-6c06-48df-a742-d9c0484ab526-ssh-key\") pod \"a6724a5b-6c06-48df-a742-d9c0484ab526\" (UID: \"a6724a5b-6c06-48df-a742-d9c0484ab526\") " Dec 03 11:26:33 crc kubenswrapper[4646]: I1203 11:26:33.852620 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6724a5b-6c06-48df-a742-d9c0484ab526-kube-api-access-hvblg" (OuterVolumeSpecName: "kube-api-access-hvblg") pod "a6724a5b-6c06-48df-a742-d9c0484ab526" (UID: "a6724a5b-6c06-48df-a742-d9c0484ab526"). InnerVolumeSpecName "kube-api-access-hvblg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:26:33 crc kubenswrapper[4646]: I1203 11:26:33.864815 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6724a5b-6c06-48df-a742-d9c0484ab526-inventory" (OuterVolumeSpecName: "inventory") pod "a6724a5b-6c06-48df-a742-d9c0484ab526" (UID: "a6724a5b-6c06-48df-a742-d9c0484ab526"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:26:33 crc kubenswrapper[4646]: I1203 11:26:33.877509 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6724a5b-6c06-48df-a742-d9c0484ab526-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a6724a5b-6c06-48df-a742-d9c0484ab526" (UID: "a6724a5b-6c06-48df-a742-d9c0484ab526"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:26:33 crc kubenswrapper[4646]: I1203 11:26:33.940758 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvblg\" (UniqueName: \"kubernetes.io/projected/a6724a5b-6c06-48df-a742-d9c0484ab526-kube-api-access-hvblg\") on node \"crc\" DevicePath \"\"" Dec 03 11:26:33 crc kubenswrapper[4646]: I1203 11:26:33.940791 4646 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6724a5b-6c06-48df-a742-d9c0484ab526-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 11:26:33 crc kubenswrapper[4646]: I1203 11:26:33.940800 4646 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6724a5b-6c06-48df-a742-d9c0484ab526-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 11:26:34 crc kubenswrapper[4646]: I1203 11:26:34.243700 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" event={"ID":"a6724a5b-6c06-48df-a742-d9c0484ab526","Type":"ContainerDied","Data":"2686b4f8bf1a07474352ae6ad8e2e15869c678fe02d758e84738abc62c028b93"} Dec 03 11:26:34 crc kubenswrapper[4646]: I1203 11:26:34.244036 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2686b4f8bf1a07474352ae6ad8e2e15869c678fe02d758e84738abc62c028b93" Dec 03 11:26:34 crc kubenswrapper[4646]: I1203 11:26:34.243738 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9" Dec 03 11:26:43 crc kubenswrapper[4646]: I1203 11:26:43.853689 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:26:43 crc kubenswrapper[4646]: E1203 11:26:43.854401 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:26:55 crc kubenswrapper[4646]: I1203 11:26:55.856702 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:26:55 crc kubenswrapper[4646]: E1203 11:26:55.857621 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:27:09 crc kubenswrapper[4646]: I1203 11:27:09.848634 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:27:10 crc kubenswrapper[4646]: I1203 11:27:10.576104 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerStarted","Data":"52c3eac8f06b635ff3f5d9230f530cf2d7c87e4c968e13300ddfce7f0bba1dcb"} Dec 03 11:29:25 crc kubenswrapper[4646]: I1203 11:29:25.964060 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:29:25 crc kubenswrapper[4646]: I1203 11:29:25.964585 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:29:55 crc kubenswrapper[4646]: I1203 11:29:55.963996 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:29:55 crc kubenswrapper[4646]: I1203 11:29:55.964489 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.200535 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg"] Dec 03 11:30:00 crc kubenswrapper[4646]: E1203 11:30:00.201511 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6724a5b-6c06-48df-a742-d9c0484ab526" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.201529 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6724a5b-6c06-48df-a742-d9c0484ab526" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.201730 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6724a5b-6c06-48df-a742-d9c0484ab526" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.202380 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.209960 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.210213 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.253588 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg"] Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.383125 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/612a2821-0fb9-49e6-8204-c7a3c013c3f4-config-volume\") pod \"collect-profiles-29412690-dgsmg\" (UID: \"612a2821-0fb9-49e6-8204-c7a3c013c3f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.383269 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6rk9\" (UniqueName: \"kubernetes.io/projected/612a2821-0fb9-49e6-8204-c7a3c013c3f4-kube-api-access-v6rk9\") pod \"collect-profiles-29412690-dgsmg\" (UID: \"612a2821-0fb9-49e6-8204-c7a3c013c3f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.383384 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/612a2821-0fb9-49e6-8204-c7a3c013c3f4-secret-volume\") pod \"collect-profiles-29412690-dgsmg\" (UID: \"612a2821-0fb9-49e6-8204-c7a3c013c3f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.484548 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/612a2821-0fb9-49e6-8204-c7a3c013c3f4-secret-volume\") pod \"collect-profiles-29412690-dgsmg\" (UID: \"612a2821-0fb9-49e6-8204-c7a3c013c3f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.484609 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/612a2821-0fb9-49e6-8204-c7a3c013c3f4-config-volume\") pod \"collect-profiles-29412690-dgsmg\" (UID: \"612a2821-0fb9-49e6-8204-c7a3c013c3f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.484695 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6rk9\" (UniqueName: \"kubernetes.io/projected/612a2821-0fb9-49e6-8204-c7a3c013c3f4-kube-api-access-v6rk9\") pod \"collect-profiles-29412690-dgsmg\" (UID: \"612a2821-0fb9-49e6-8204-c7a3c013c3f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.486016 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/612a2821-0fb9-49e6-8204-c7a3c013c3f4-config-volume\") pod \"collect-profiles-29412690-dgsmg\" (UID: \"612a2821-0fb9-49e6-8204-c7a3c013c3f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.503363 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/612a2821-0fb9-49e6-8204-c7a3c013c3f4-secret-volume\") pod \"collect-profiles-29412690-dgsmg\" (UID: \"612a2821-0fb9-49e6-8204-c7a3c013c3f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.504378 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6rk9\" (UniqueName: \"kubernetes.io/projected/612a2821-0fb9-49e6-8204-c7a3c013c3f4-kube-api-access-v6rk9\") pod \"collect-profiles-29412690-dgsmg\" (UID: \"612a2821-0fb9-49e6-8204-c7a3c013c3f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.539682 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" Dec 03 11:30:00 crc kubenswrapper[4646]: I1203 11:30:00.984509 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg"] Dec 03 11:30:01 crc kubenswrapper[4646]: I1203 11:30:01.978096 4646 generic.go:334] "Generic (PLEG): container finished" podID="612a2821-0fb9-49e6-8204-c7a3c013c3f4" containerID="28cf314a2234686931f6fd37fa69ba3bd968baf3a6776d1f73a097dd152e1cac" exitCode=0 Dec 03 11:30:01 crc kubenswrapper[4646]: I1203 11:30:01.978163 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" event={"ID":"612a2821-0fb9-49e6-8204-c7a3c013c3f4","Type":"ContainerDied","Data":"28cf314a2234686931f6fd37fa69ba3bd968baf3a6776d1f73a097dd152e1cac"} Dec 03 11:30:01 crc kubenswrapper[4646]: I1203 11:30:01.978720 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" event={"ID":"612a2821-0fb9-49e6-8204-c7a3c013c3f4","Type":"ContainerStarted","Data":"650eab7a709ae1ef27b6b018c242e7045e585d0ad0c81a9db85ea643fa208581"} Dec 03 11:30:03 crc kubenswrapper[4646]: I1203 11:30:03.337035 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" Dec 03 11:30:03 crc kubenswrapper[4646]: I1203 11:30:03.440997 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6rk9\" (UniqueName: \"kubernetes.io/projected/612a2821-0fb9-49e6-8204-c7a3c013c3f4-kube-api-access-v6rk9\") pod \"612a2821-0fb9-49e6-8204-c7a3c013c3f4\" (UID: \"612a2821-0fb9-49e6-8204-c7a3c013c3f4\") " Dec 03 11:30:03 crc kubenswrapper[4646]: I1203 11:30:03.441200 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/612a2821-0fb9-49e6-8204-c7a3c013c3f4-config-volume\") pod \"612a2821-0fb9-49e6-8204-c7a3c013c3f4\" (UID: \"612a2821-0fb9-49e6-8204-c7a3c013c3f4\") " Dec 03 11:30:03 crc kubenswrapper[4646]: I1203 11:30:03.441279 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/612a2821-0fb9-49e6-8204-c7a3c013c3f4-secret-volume\") pod \"612a2821-0fb9-49e6-8204-c7a3c013c3f4\" (UID: \"612a2821-0fb9-49e6-8204-c7a3c013c3f4\") " Dec 03 11:30:03 crc kubenswrapper[4646]: I1203 11:30:03.448042 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/612a2821-0fb9-49e6-8204-c7a3c013c3f4-config-volume" (OuterVolumeSpecName: "config-volume") pod "612a2821-0fb9-49e6-8204-c7a3c013c3f4" (UID: "612a2821-0fb9-49e6-8204-c7a3c013c3f4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 11:30:03 crc kubenswrapper[4646]: I1203 11:30:03.448920 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/612a2821-0fb9-49e6-8204-c7a3c013c3f4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "612a2821-0fb9-49e6-8204-c7a3c013c3f4" (UID: "612a2821-0fb9-49e6-8204-c7a3c013c3f4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 11:30:03 crc kubenswrapper[4646]: I1203 11:30:03.450106 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/612a2821-0fb9-49e6-8204-c7a3c013c3f4-kube-api-access-v6rk9" (OuterVolumeSpecName: "kube-api-access-v6rk9") pod "612a2821-0fb9-49e6-8204-c7a3c013c3f4" (UID: "612a2821-0fb9-49e6-8204-c7a3c013c3f4"). InnerVolumeSpecName "kube-api-access-v6rk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:30:03 crc kubenswrapper[4646]: I1203 11:30:03.543728 4646 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/612a2821-0fb9-49e6-8204-c7a3c013c3f4-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 11:30:03 crc kubenswrapper[4646]: I1203 11:30:03.543766 4646 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/612a2821-0fb9-49e6-8204-c7a3c013c3f4-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 11:30:03 crc kubenswrapper[4646]: I1203 11:30:03.543777 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6rk9\" (UniqueName: \"kubernetes.io/projected/612a2821-0fb9-49e6-8204-c7a3c013c3f4-kube-api-access-v6rk9\") on node \"crc\" DevicePath \"\"" Dec 03 11:30:03 crc kubenswrapper[4646]: I1203 11:30:03.995571 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" event={"ID":"612a2821-0fb9-49e6-8204-c7a3c013c3f4","Type":"ContainerDied","Data":"650eab7a709ae1ef27b6b018c242e7045e585d0ad0c81a9db85ea643fa208581"} Dec 03 11:30:03 crc kubenswrapper[4646]: I1203 11:30:03.995672 4646 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="650eab7a709ae1ef27b6b018c242e7045e585d0ad0c81a9db85ea643fa208581" Dec 03 11:30:03 crc kubenswrapper[4646]: I1203 11:30:03.995615 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412690-dgsmg" Dec 03 11:30:04 crc kubenswrapper[4646]: I1203 11:30:04.414127 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t"] Dec 03 11:30:04 crc kubenswrapper[4646]: I1203 11:30:04.455911 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412645-6w78t"] Dec 03 11:30:05 crc kubenswrapper[4646]: I1203 11:30:05.862698 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8e77566-f3a2-4a10-84b5-4e6078bc6f23" path="/var/lib/kubelet/pods/d8e77566-f3a2-4a10-84b5-4e6078bc6f23/volumes" Dec 03 11:30:25 crc kubenswrapper[4646]: I1203 11:30:25.964383 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:30:25 crc kubenswrapper[4646]: I1203 11:30:25.965033 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:30:25 crc kubenswrapper[4646]: I1203 11:30:25.965089 4646 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 11:30:25 crc kubenswrapper[4646]: I1203 11:30:25.966023 4646 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"52c3eac8f06b635ff3f5d9230f530cf2d7c87e4c968e13300ddfce7f0bba1dcb"} pod="openshift-machine-config-operator/machine-config-daemon-pndzt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 11:30:25 crc kubenswrapper[4646]: I1203 11:30:25.966195 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" containerID="cri-o://52c3eac8f06b635ff3f5d9230f530cf2d7c87e4c968e13300ddfce7f0bba1dcb" gracePeriod=600 Dec 03 11:30:26 crc kubenswrapper[4646]: I1203 11:30:26.177823 4646 generic.go:334] "Generic (PLEG): container finished" podID="9695d947-fc43-4625-86af-a1a3e177bb27" containerID="52c3eac8f06b635ff3f5d9230f530cf2d7c87e4c968e13300ddfce7f0bba1dcb" exitCode=0 Dec 03 11:30:26 crc kubenswrapper[4646]: I1203 11:30:26.177875 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerDied","Data":"52c3eac8f06b635ff3f5d9230f530cf2d7c87e4c968e13300ddfce7f0bba1dcb"} Dec 03 11:30:26 crc kubenswrapper[4646]: I1203 11:30:26.177932 4646 scope.go:117] "RemoveContainer" containerID="1f937994c98bfd7f39c046e5bb698e3f6078edf08c3919254afbd4e4acb4c30e" Dec 03 11:30:27 crc kubenswrapper[4646]: I1203 11:30:27.186725 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerStarted","Data":"19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29"} Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.117713 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7cz2g"] Dec 03 11:30:30 crc kubenswrapper[4646]: E1203 11:30:30.118828 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="612a2821-0fb9-49e6-8204-c7a3c013c3f4" containerName="collect-profiles" Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.118852 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="612a2821-0fb9-49e6-8204-c7a3c013c3f4" containerName="collect-profiles" Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.119078 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="612a2821-0fb9-49e6-8204-c7a3c013c3f4" containerName="collect-profiles" Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.127300 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.132640 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cz2g"] Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.230141 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48628bae-9010-4f11-90e1-f0c78be31a57-catalog-content\") pod \"redhat-marketplace-7cz2g\" (UID: \"48628bae-9010-4f11-90e1-f0c78be31a57\") " pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.230691 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8r8k\" (UniqueName: \"kubernetes.io/projected/48628bae-9010-4f11-90e1-f0c78be31a57-kube-api-access-v8r8k\") pod \"redhat-marketplace-7cz2g\" (UID: \"48628bae-9010-4f11-90e1-f0c78be31a57\") " pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.230768 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48628bae-9010-4f11-90e1-f0c78be31a57-utilities\") pod \"redhat-marketplace-7cz2g\" (UID: \"48628bae-9010-4f11-90e1-f0c78be31a57\") " pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.332876 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48628bae-9010-4f11-90e1-f0c78be31a57-catalog-content\") pod \"redhat-marketplace-7cz2g\" (UID: \"48628bae-9010-4f11-90e1-f0c78be31a57\") " pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.333006 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8r8k\" (UniqueName: \"kubernetes.io/projected/48628bae-9010-4f11-90e1-f0c78be31a57-kube-api-access-v8r8k\") pod \"redhat-marketplace-7cz2g\" (UID: \"48628bae-9010-4f11-90e1-f0c78be31a57\") " pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.333047 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48628bae-9010-4f11-90e1-f0c78be31a57-utilities\") pod \"redhat-marketplace-7cz2g\" (UID: \"48628bae-9010-4f11-90e1-f0c78be31a57\") " pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.333539 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48628bae-9010-4f11-90e1-f0c78be31a57-utilities\") pod \"redhat-marketplace-7cz2g\" (UID: \"48628bae-9010-4f11-90e1-f0c78be31a57\") " pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.333767 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48628bae-9010-4f11-90e1-f0c78be31a57-catalog-content\") pod \"redhat-marketplace-7cz2g\" (UID: \"48628bae-9010-4f11-90e1-f0c78be31a57\") " pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.352651 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8r8k\" (UniqueName: \"kubernetes.io/projected/48628bae-9010-4f11-90e1-f0c78be31a57-kube-api-access-v8r8k\") pod \"redhat-marketplace-7cz2g\" (UID: \"48628bae-9010-4f11-90e1-f0c78be31a57\") " pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.453780 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:30 crc kubenswrapper[4646]: I1203 11:30:30.978949 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cz2g"] Dec 03 11:30:31 crc kubenswrapper[4646]: I1203 11:30:31.236606 4646 generic.go:334] "Generic (PLEG): container finished" podID="48628bae-9010-4f11-90e1-f0c78be31a57" containerID="fe80c0a47f12e7ba7e233eb24b5cbceb241085125fcffc837fac8aaa043e5bcb" exitCode=0 Dec 03 11:30:31 crc kubenswrapper[4646]: I1203 11:30:31.236663 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cz2g" event={"ID":"48628bae-9010-4f11-90e1-f0c78be31a57","Type":"ContainerDied","Data":"fe80c0a47f12e7ba7e233eb24b5cbceb241085125fcffc837fac8aaa043e5bcb"} Dec 03 11:30:31 crc kubenswrapper[4646]: I1203 11:30:31.236714 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cz2g" event={"ID":"48628bae-9010-4f11-90e1-f0c78be31a57","Type":"ContainerStarted","Data":"562fc82c8162ef042336426623e0782080bd463792c747765ca33eec27b63de0"} Dec 03 11:30:32 crc kubenswrapper[4646]: I1203 11:30:32.246756 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cz2g" event={"ID":"48628bae-9010-4f11-90e1-f0c78be31a57","Type":"ContainerStarted","Data":"c6b36ec3fa25afed48c8fda71d482abd3d922dd7ee954c71c9cdb7038d2a9b91"} Dec 03 11:30:33 crc kubenswrapper[4646]: I1203 11:30:33.261822 4646 generic.go:334] "Generic (PLEG): container finished" podID="48628bae-9010-4f11-90e1-f0c78be31a57" containerID="c6b36ec3fa25afed48c8fda71d482abd3d922dd7ee954c71c9cdb7038d2a9b91" exitCode=0 Dec 03 11:30:33 crc kubenswrapper[4646]: I1203 11:30:33.262006 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cz2g" event={"ID":"48628bae-9010-4f11-90e1-f0c78be31a57","Type":"ContainerDied","Data":"c6b36ec3fa25afed48c8fda71d482abd3d922dd7ee954c71c9cdb7038d2a9b91"} Dec 03 11:30:34 crc kubenswrapper[4646]: I1203 11:30:34.274216 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cz2g" event={"ID":"48628bae-9010-4f11-90e1-f0c78be31a57","Type":"ContainerStarted","Data":"75a7f15a5ff38f1ea4c76fcf0dad0ee2ebb1400f44f4d3a559c83f75e8fc3ed9"} Dec 03 11:30:34 crc kubenswrapper[4646]: I1203 11:30:34.299731 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7cz2g" podStartSLOduration=1.890234097 podStartE2EDuration="4.299711231s" podCreationTimestamp="2025-12-03 11:30:30 +0000 UTC" firstStartedPulling="2025-12-03 11:30:31.238828965 +0000 UTC m=+2207.701885100" lastFinishedPulling="2025-12-03 11:30:33.648306089 +0000 UTC m=+2210.111362234" observedRunningTime="2025-12-03 11:30:34.293243627 +0000 UTC m=+2210.756299762" watchObservedRunningTime="2025-12-03 11:30:34.299711231 +0000 UTC m=+2210.762767356" Dec 03 11:30:40 crc kubenswrapper[4646]: I1203 11:30:40.454760 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:40 crc kubenswrapper[4646]: I1203 11:30:40.455305 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:40 crc kubenswrapper[4646]: I1203 11:30:40.507294 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:41 crc kubenswrapper[4646]: I1203 11:30:41.393695 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:41 crc kubenswrapper[4646]: I1203 11:30:41.436853 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cz2g"] Dec 03 11:30:43 crc kubenswrapper[4646]: I1203 11:30:43.365113 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7cz2g" podUID="48628bae-9010-4f11-90e1-f0c78be31a57" containerName="registry-server" containerID="cri-o://75a7f15a5ff38f1ea4c76fcf0dad0ee2ebb1400f44f4d3a559c83f75e8fc3ed9" gracePeriod=2 Dec 03 11:30:43 crc kubenswrapper[4646]: I1203 11:30:43.750583 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:43 crc kubenswrapper[4646]: I1203 11:30:43.903856 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8r8k\" (UniqueName: \"kubernetes.io/projected/48628bae-9010-4f11-90e1-f0c78be31a57-kube-api-access-v8r8k\") pod \"48628bae-9010-4f11-90e1-f0c78be31a57\" (UID: \"48628bae-9010-4f11-90e1-f0c78be31a57\") " Dec 03 11:30:43 crc kubenswrapper[4646]: I1203 11:30:43.903904 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48628bae-9010-4f11-90e1-f0c78be31a57-catalog-content\") pod \"48628bae-9010-4f11-90e1-f0c78be31a57\" (UID: \"48628bae-9010-4f11-90e1-f0c78be31a57\") " Dec 03 11:30:43 crc kubenswrapper[4646]: I1203 11:30:43.904199 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48628bae-9010-4f11-90e1-f0c78be31a57-utilities\") pod \"48628bae-9010-4f11-90e1-f0c78be31a57\" (UID: \"48628bae-9010-4f11-90e1-f0c78be31a57\") " Dec 03 11:30:43 crc kubenswrapper[4646]: I1203 11:30:43.905410 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48628bae-9010-4f11-90e1-f0c78be31a57-utilities" (OuterVolumeSpecName: "utilities") pod "48628bae-9010-4f11-90e1-f0c78be31a57" (UID: "48628bae-9010-4f11-90e1-f0c78be31a57"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:30:43 crc kubenswrapper[4646]: I1203 11:30:43.909304 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48628bae-9010-4f11-90e1-f0c78be31a57-kube-api-access-v8r8k" (OuterVolumeSpecName: "kube-api-access-v8r8k") pod "48628bae-9010-4f11-90e1-f0c78be31a57" (UID: "48628bae-9010-4f11-90e1-f0c78be31a57"). InnerVolumeSpecName "kube-api-access-v8r8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:30:43 crc kubenswrapper[4646]: I1203 11:30:43.924456 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48628bae-9010-4f11-90e1-f0c78be31a57-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48628bae-9010-4f11-90e1-f0c78be31a57" (UID: "48628bae-9010-4f11-90e1-f0c78be31a57"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.006915 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8r8k\" (UniqueName: \"kubernetes.io/projected/48628bae-9010-4f11-90e1-f0c78be31a57-kube-api-access-v8r8k\") on node \"crc\" DevicePath \"\"" Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.006954 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48628bae-9010-4f11-90e1-f0c78be31a57-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.006967 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48628bae-9010-4f11-90e1-f0c78be31a57-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.376580 4646 generic.go:334] "Generic (PLEG): container finished" podID="48628bae-9010-4f11-90e1-f0c78be31a57" containerID="75a7f15a5ff38f1ea4c76fcf0dad0ee2ebb1400f44f4d3a559c83f75e8fc3ed9" exitCode=0 Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.376641 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cz2g" Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.376660 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cz2g" event={"ID":"48628bae-9010-4f11-90e1-f0c78be31a57","Type":"ContainerDied","Data":"75a7f15a5ff38f1ea4c76fcf0dad0ee2ebb1400f44f4d3a559c83f75e8fc3ed9"} Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.377062 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cz2g" event={"ID":"48628bae-9010-4f11-90e1-f0c78be31a57","Type":"ContainerDied","Data":"562fc82c8162ef042336426623e0782080bd463792c747765ca33eec27b63de0"} Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.377085 4646 scope.go:117] "RemoveContainer" containerID="75a7f15a5ff38f1ea4c76fcf0dad0ee2ebb1400f44f4d3a559c83f75e8fc3ed9" Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.397693 4646 scope.go:117] "RemoveContainer" containerID="c6b36ec3fa25afed48c8fda71d482abd3d922dd7ee954c71c9cdb7038d2a9b91" Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.444201 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cz2g"] Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.444595 4646 scope.go:117] "RemoveContainer" containerID="fe80c0a47f12e7ba7e233eb24b5cbceb241085125fcffc837fac8aaa043e5bcb" Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.454823 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cz2g"] Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.474825 4646 scope.go:117] "RemoveContainer" containerID="75a7f15a5ff38f1ea4c76fcf0dad0ee2ebb1400f44f4d3a559c83f75e8fc3ed9" Dec 03 11:30:44 crc kubenswrapper[4646]: E1203 11:30:44.477084 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75a7f15a5ff38f1ea4c76fcf0dad0ee2ebb1400f44f4d3a559c83f75e8fc3ed9\": container with ID starting with 75a7f15a5ff38f1ea4c76fcf0dad0ee2ebb1400f44f4d3a559c83f75e8fc3ed9 not found: ID does not exist" containerID="75a7f15a5ff38f1ea4c76fcf0dad0ee2ebb1400f44f4d3a559c83f75e8fc3ed9" Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.477131 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75a7f15a5ff38f1ea4c76fcf0dad0ee2ebb1400f44f4d3a559c83f75e8fc3ed9"} err="failed to get container status \"75a7f15a5ff38f1ea4c76fcf0dad0ee2ebb1400f44f4d3a559c83f75e8fc3ed9\": rpc error: code = NotFound desc = could not find container \"75a7f15a5ff38f1ea4c76fcf0dad0ee2ebb1400f44f4d3a559c83f75e8fc3ed9\": container with ID starting with 75a7f15a5ff38f1ea4c76fcf0dad0ee2ebb1400f44f4d3a559c83f75e8fc3ed9 not found: ID does not exist" Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.477164 4646 scope.go:117] "RemoveContainer" containerID="c6b36ec3fa25afed48c8fda71d482abd3d922dd7ee954c71c9cdb7038d2a9b91" Dec 03 11:30:44 crc kubenswrapper[4646]: E1203 11:30:44.477607 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6b36ec3fa25afed48c8fda71d482abd3d922dd7ee954c71c9cdb7038d2a9b91\": container with ID starting with c6b36ec3fa25afed48c8fda71d482abd3d922dd7ee954c71c9cdb7038d2a9b91 not found: ID does not exist" containerID="c6b36ec3fa25afed48c8fda71d482abd3d922dd7ee954c71c9cdb7038d2a9b91" Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.477693 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b36ec3fa25afed48c8fda71d482abd3d922dd7ee954c71c9cdb7038d2a9b91"} err="failed to get container status \"c6b36ec3fa25afed48c8fda71d482abd3d922dd7ee954c71c9cdb7038d2a9b91\": rpc error: code = NotFound desc = could not find container \"c6b36ec3fa25afed48c8fda71d482abd3d922dd7ee954c71c9cdb7038d2a9b91\": container with ID starting with c6b36ec3fa25afed48c8fda71d482abd3d922dd7ee954c71c9cdb7038d2a9b91 not found: ID does not exist" Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.477762 4646 scope.go:117] "RemoveContainer" containerID="fe80c0a47f12e7ba7e233eb24b5cbceb241085125fcffc837fac8aaa043e5bcb" Dec 03 11:30:44 crc kubenswrapper[4646]: E1203 11:30:44.478177 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe80c0a47f12e7ba7e233eb24b5cbceb241085125fcffc837fac8aaa043e5bcb\": container with ID starting with fe80c0a47f12e7ba7e233eb24b5cbceb241085125fcffc837fac8aaa043e5bcb not found: ID does not exist" containerID="fe80c0a47f12e7ba7e233eb24b5cbceb241085125fcffc837fac8aaa043e5bcb" Dec 03 11:30:44 crc kubenswrapper[4646]: I1203 11:30:44.478271 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe80c0a47f12e7ba7e233eb24b5cbceb241085125fcffc837fac8aaa043e5bcb"} err="failed to get container status \"fe80c0a47f12e7ba7e233eb24b5cbceb241085125fcffc837fac8aaa043e5bcb\": rpc error: code = NotFound desc = could not find container \"fe80c0a47f12e7ba7e233eb24b5cbceb241085125fcffc837fac8aaa043e5bcb\": container with ID starting with fe80c0a47f12e7ba7e233eb24b5cbceb241085125fcffc837fac8aaa043e5bcb not found: ID does not exist" Dec 03 11:30:45 crc kubenswrapper[4646]: I1203 11:30:45.860904 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48628bae-9010-4f11-90e1-f0c78be31a57" path="/var/lib/kubelet/pods/48628bae-9010-4f11-90e1-f0c78be31a57/volumes" Dec 03 11:31:05 crc kubenswrapper[4646]: I1203 11:31:05.827528 4646 scope.go:117] "RemoveContainer" containerID="011d2186daa42776238b6189b1866584dcfdc7148abe0731e3b9c1021271b63d" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.563106 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vgmc6"] Dec 03 11:31:21 crc kubenswrapper[4646]: E1203 11:31:21.564456 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48628bae-9010-4f11-90e1-f0c78be31a57" containerName="extract-utilities" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.564496 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="48628bae-9010-4f11-90e1-f0c78be31a57" containerName="extract-utilities" Dec 03 11:31:21 crc kubenswrapper[4646]: E1203 11:31:21.564520 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48628bae-9010-4f11-90e1-f0c78be31a57" containerName="extract-content" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.564533 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="48628bae-9010-4f11-90e1-f0c78be31a57" containerName="extract-content" Dec 03 11:31:21 crc kubenswrapper[4646]: E1203 11:31:21.564605 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48628bae-9010-4f11-90e1-f0c78be31a57" containerName="registry-server" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.564618 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="48628bae-9010-4f11-90e1-f0c78be31a57" containerName="registry-server" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.564992 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="48628bae-9010-4f11-90e1-f0c78be31a57" containerName="registry-server" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.567619 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.578716 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vgmc6"] Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.590770 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3fec42e-082c-406e-bf52-a6bf399c5c8b-catalog-content\") pod \"community-operators-vgmc6\" (UID: \"b3fec42e-082c-406e-bf52-a6bf399c5c8b\") " pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.590821 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52bhz\" (UniqueName: \"kubernetes.io/projected/b3fec42e-082c-406e-bf52-a6bf399c5c8b-kube-api-access-52bhz\") pod \"community-operators-vgmc6\" (UID: \"b3fec42e-082c-406e-bf52-a6bf399c5c8b\") " pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.590853 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3fec42e-082c-406e-bf52-a6bf399c5c8b-utilities\") pod \"community-operators-vgmc6\" (UID: \"b3fec42e-082c-406e-bf52-a6bf399c5c8b\") " pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.692486 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3fec42e-082c-406e-bf52-a6bf399c5c8b-catalog-content\") pod \"community-operators-vgmc6\" (UID: \"b3fec42e-082c-406e-bf52-a6bf399c5c8b\") " pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.692603 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52bhz\" (UniqueName: \"kubernetes.io/projected/b3fec42e-082c-406e-bf52-a6bf399c5c8b-kube-api-access-52bhz\") pod \"community-operators-vgmc6\" (UID: \"b3fec42e-082c-406e-bf52-a6bf399c5c8b\") " pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.692653 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3fec42e-082c-406e-bf52-a6bf399c5c8b-utilities\") pod \"community-operators-vgmc6\" (UID: \"b3fec42e-082c-406e-bf52-a6bf399c5c8b\") " pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.693014 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3fec42e-082c-406e-bf52-a6bf399c5c8b-catalog-content\") pod \"community-operators-vgmc6\" (UID: \"b3fec42e-082c-406e-bf52-a6bf399c5c8b\") " pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.693080 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3fec42e-082c-406e-bf52-a6bf399c5c8b-utilities\") pod \"community-operators-vgmc6\" (UID: \"b3fec42e-082c-406e-bf52-a6bf399c5c8b\") " pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.713503 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52bhz\" (UniqueName: \"kubernetes.io/projected/b3fec42e-082c-406e-bf52-a6bf399c5c8b-kube-api-access-52bhz\") pod \"community-operators-vgmc6\" (UID: \"b3fec42e-082c-406e-bf52-a6bf399c5c8b\") " pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:21 crc kubenswrapper[4646]: I1203 11:31:21.888200 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:22 crc kubenswrapper[4646]: I1203 11:31:22.496871 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vgmc6"] Dec 03 11:31:22 crc kubenswrapper[4646]: I1203 11:31:22.697155 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgmc6" event={"ID":"b3fec42e-082c-406e-bf52-a6bf399c5c8b","Type":"ContainerStarted","Data":"9aca43448ec9eefbf882a1c0cbac88e4a7e2d96af0de3e47f2626b81946cd08b"} Dec 03 11:31:22 crc kubenswrapper[4646]: I1203 11:31:22.697638 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgmc6" event={"ID":"b3fec42e-082c-406e-bf52-a6bf399c5c8b","Type":"ContainerStarted","Data":"a1c3881589581274d1501854856383a80e04b81b4200cda5d711d8007ecde6d3"} Dec 03 11:31:23 crc kubenswrapper[4646]: I1203 11:31:23.707169 4646 generic.go:334] "Generic (PLEG): container finished" podID="b3fec42e-082c-406e-bf52-a6bf399c5c8b" containerID="9aca43448ec9eefbf882a1c0cbac88e4a7e2d96af0de3e47f2626b81946cd08b" exitCode=0 Dec 03 11:31:23 crc kubenswrapper[4646]: I1203 11:31:23.707287 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgmc6" event={"ID":"b3fec42e-082c-406e-bf52-a6bf399c5c8b","Type":"ContainerDied","Data":"9aca43448ec9eefbf882a1c0cbac88e4a7e2d96af0de3e47f2626b81946cd08b"} Dec 03 11:31:23 crc kubenswrapper[4646]: I1203 11:31:23.710512 4646 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 11:31:25 crc kubenswrapper[4646]: I1203 11:31:25.728448 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgmc6" event={"ID":"b3fec42e-082c-406e-bf52-a6bf399c5c8b","Type":"ContainerStarted","Data":"5903c633821f225c10256df24acd67cd3495c6515af6029296408c041a88a53b"} Dec 03 11:31:26 crc kubenswrapper[4646]: I1203 11:31:26.738310 4646 generic.go:334] "Generic (PLEG): container finished" podID="b3fec42e-082c-406e-bf52-a6bf399c5c8b" containerID="5903c633821f225c10256df24acd67cd3495c6515af6029296408c041a88a53b" exitCode=0 Dec 03 11:31:26 crc kubenswrapper[4646]: I1203 11:31:26.738427 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgmc6" event={"ID":"b3fec42e-082c-406e-bf52-a6bf399c5c8b","Type":"ContainerDied","Data":"5903c633821f225c10256df24acd67cd3495c6515af6029296408c041a88a53b"} Dec 03 11:31:27 crc kubenswrapper[4646]: I1203 11:31:27.749096 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgmc6" event={"ID":"b3fec42e-082c-406e-bf52-a6bf399c5c8b","Type":"ContainerStarted","Data":"ed3de6676454924ea709ed77e4d539aadff773885b4bac3facc9de7625ce10c8"} Dec 03 11:31:27 crc kubenswrapper[4646]: I1203 11:31:27.772003 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vgmc6" podStartSLOduration=3.377300549 podStartE2EDuration="6.771983762s" podCreationTimestamp="2025-12-03 11:31:21 +0000 UTC" firstStartedPulling="2025-12-03 11:31:23.710242412 +0000 UTC m=+2260.173298537" lastFinishedPulling="2025-12-03 11:31:27.104925615 +0000 UTC m=+2263.567981750" observedRunningTime="2025-12-03 11:31:27.764381736 +0000 UTC m=+2264.227437891" watchObservedRunningTime="2025-12-03 11:31:27.771983762 +0000 UTC m=+2264.235039897" Dec 03 11:31:31 crc kubenswrapper[4646]: I1203 11:31:31.888036 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:31 crc kubenswrapper[4646]: I1203 11:31:31.888582 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:31 crc kubenswrapper[4646]: I1203 11:31:31.939922 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:32 crc kubenswrapper[4646]: I1203 11:31:32.850653 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:32 crc kubenswrapper[4646]: I1203 11:31:32.894046 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vgmc6"] Dec 03 11:31:34 crc kubenswrapper[4646]: I1203 11:31:34.823218 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vgmc6" podUID="b3fec42e-082c-406e-bf52-a6bf399c5c8b" containerName="registry-server" containerID="cri-o://ed3de6676454924ea709ed77e4d539aadff773885b4bac3facc9de7625ce10c8" gracePeriod=2 Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.261251 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.363865 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52bhz\" (UniqueName: \"kubernetes.io/projected/b3fec42e-082c-406e-bf52-a6bf399c5c8b-kube-api-access-52bhz\") pod \"b3fec42e-082c-406e-bf52-a6bf399c5c8b\" (UID: \"b3fec42e-082c-406e-bf52-a6bf399c5c8b\") " Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.364193 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3fec42e-082c-406e-bf52-a6bf399c5c8b-catalog-content\") pod \"b3fec42e-082c-406e-bf52-a6bf399c5c8b\" (UID: \"b3fec42e-082c-406e-bf52-a6bf399c5c8b\") " Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.364313 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3fec42e-082c-406e-bf52-a6bf399c5c8b-utilities\") pod \"b3fec42e-082c-406e-bf52-a6bf399c5c8b\" (UID: \"b3fec42e-082c-406e-bf52-a6bf399c5c8b\") " Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.376548 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3fec42e-082c-406e-bf52-a6bf399c5c8b-utilities" (OuterVolumeSpecName: "utilities") pod "b3fec42e-082c-406e-bf52-a6bf399c5c8b" (UID: "b3fec42e-082c-406e-bf52-a6bf399c5c8b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.382769 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3fec42e-082c-406e-bf52-a6bf399c5c8b-kube-api-access-52bhz" (OuterVolumeSpecName: "kube-api-access-52bhz") pod "b3fec42e-082c-406e-bf52-a6bf399c5c8b" (UID: "b3fec42e-082c-406e-bf52-a6bf399c5c8b"). InnerVolumeSpecName "kube-api-access-52bhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.428142 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3fec42e-082c-406e-bf52-a6bf399c5c8b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3fec42e-082c-406e-bf52-a6bf399c5c8b" (UID: "b3fec42e-082c-406e-bf52-a6bf399c5c8b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.467695 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52bhz\" (UniqueName: \"kubernetes.io/projected/b3fec42e-082c-406e-bf52-a6bf399c5c8b-kube-api-access-52bhz\") on node \"crc\" DevicePath \"\"" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.467742 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3fec42e-082c-406e-bf52-a6bf399c5c8b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.467754 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3fec42e-082c-406e-bf52-a6bf399c5c8b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.833159 4646 generic.go:334] "Generic (PLEG): container finished" podID="b3fec42e-082c-406e-bf52-a6bf399c5c8b" containerID="ed3de6676454924ea709ed77e4d539aadff773885b4bac3facc9de7625ce10c8" exitCode=0 Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.833203 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgmc6" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.833216 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgmc6" event={"ID":"b3fec42e-082c-406e-bf52-a6bf399c5c8b","Type":"ContainerDied","Data":"ed3de6676454924ea709ed77e4d539aadff773885b4bac3facc9de7625ce10c8"} Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.833256 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgmc6" event={"ID":"b3fec42e-082c-406e-bf52-a6bf399c5c8b","Type":"ContainerDied","Data":"a1c3881589581274d1501854856383a80e04b81b4200cda5d711d8007ecde6d3"} Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.833276 4646 scope.go:117] "RemoveContainer" containerID="ed3de6676454924ea709ed77e4d539aadff773885b4bac3facc9de7625ce10c8" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.860810 4646 scope.go:117] "RemoveContainer" containerID="5903c633821f225c10256df24acd67cd3495c6515af6029296408c041a88a53b" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.886551 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vgmc6"] Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.894617 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vgmc6"] Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.906214 4646 scope.go:117] "RemoveContainer" containerID="9aca43448ec9eefbf882a1c0cbac88e4a7e2d96af0de3e47f2626b81946cd08b" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.926618 4646 scope.go:117] "RemoveContainer" containerID="ed3de6676454924ea709ed77e4d539aadff773885b4bac3facc9de7625ce10c8" Dec 03 11:31:35 crc kubenswrapper[4646]: E1203 11:31:35.927052 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed3de6676454924ea709ed77e4d539aadff773885b4bac3facc9de7625ce10c8\": container with ID starting with ed3de6676454924ea709ed77e4d539aadff773885b4bac3facc9de7625ce10c8 not found: ID does not exist" containerID="ed3de6676454924ea709ed77e4d539aadff773885b4bac3facc9de7625ce10c8" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.927107 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed3de6676454924ea709ed77e4d539aadff773885b4bac3facc9de7625ce10c8"} err="failed to get container status \"ed3de6676454924ea709ed77e4d539aadff773885b4bac3facc9de7625ce10c8\": rpc error: code = NotFound desc = could not find container \"ed3de6676454924ea709ed77e4d539aadff773885b4bac3facc9de7625ce10c8\": container with ID starting with ed3de6676454924ea709ed77e4d539aadff773885b4bac3facc9de7625ce10c8 not found: ID does not exist" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.927141 4646 scope.go:117] "RemoveContainer" containerID="5903c633821f225c10256df24acd67cd3495c6515af6029296408c041a88a53b" Dec 03 11:31:35 crc kubenswrapper[4646]: E1203 11:31:35.927512 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5903c633821f225c10256df24acd67cd3495c6515af6029296408c041a88a53b\": container with ID starting with 5903c633821f225c10256df24acd67cd3495c6515af6029296408c041a88a53b not found: ID does not exist" containerID="5903c633821f225c10256df24acd67cd3495c6515af6029296408c041a88a53b" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.927554 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5903c633821f225c10256df24acd67cd3495c6515af6029296408c041a88a53b"} err="failed to get container status \"5903c633821f225c10256df24acd67cd3495c6515af6029296408c041a88a53b\": rpc error: code = NotFound desc = could not find container \"5903c633821f225c10256df24acd67cd3495c6515af6029296408c041a88a53b\": container with ID starting with 5903c633821f225c10256df24acd67cd3495c6515af6029296408c041a88a53b not found: ID does not exist" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.927588 4646 scope.go:117] "RemoveContainer" containerID="9aca43448ec9eefbf882a1c0cbac88e4a7e2d96af0de3e47f2626b81946cd08b" Dec 03 11:31:35 crc kubenswrapper[4646]: E1203 11:31:35.927846 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9aca43448ec9eefbf882a1c0cbac88e4a7e2d96af0de3e47f2626b81946cd08b\": container with ID starting with 9aca43448ec9eefbf882a1c0cbac88e4a7e2d96af0de3e47f2626b81946cd08b not found: ID does not exist" containerID="9aca43448ec9eefbf882a1c0cbac88e4a7e2d96af0de3e47f2626b81946cd08b" Dec 03 11:31:35 crc kubenswrapper[4646]: I1203 11:31:35.927874 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aca43448ec9eefbf882a1c0cbac88e4a7e2d96af0de3e47f2626b81946cd08b"} err="failed to get container status \"9aca43448ec9eefbf882a1c0cbac88e4a7e2d96af0de3e47f2626b81946cd08b\": rpc error: code = NotFound desc = could not find container \"9aca43448ec9eefbf882a1c0cbac88e4a7e2d96af0de3e47f2626b81946cd08b\": container with ID starting with 9aca43448ec9eefbf882a1c0cbac88e4a7e2d96af0de3e47f2626b81946cd08b not found: ID does not exist" Dec 03 11:31:37 crc kubenswrapper[4646]: I1203 11:31:37.870021 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3fec42e-082c-406e-bf52-a6bf399c5c8b" path="/var/lib/kubelet/pods/b3fec42e-082c-406e-bf52-a6bf399c5c8b/volumes" Dec 03 11:32:55 crc kubenswrapper[4646]: I1203 11:32:55.964168 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:32:55 crc kubenswrapper[4646]: I1203 11:32:55.964707 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:33:25 crc kubenswrapper[4646]: I1203 11:33:25.964690 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:33:25 crc kubenswrapper[4646]: I1203 11:33:25.965983 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.362609 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-74vwp/must-gather-fxg6j"] Dec 03 11:33:32 crc kubenswrapper[4646]: E1203 11:33:32.363678 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3fec42e-082c-406e-bf52-a6bf399c5c8b" containerName="extract-content" Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.363698 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3fec42e-082c-406e-bf52-a6bf399c5c8b" containerName="extract-content" Dec 03 11:33:32 crc kubenswrapper[4646]: E1203 11:33:32.363711 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3fec42e-082c-406e-bf52-a6bf399c5c8b" containerName="extract-utilities" Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.363719 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3fec42e-082c-406e-bf52-a6bf399c5c8b" containerName="extract-utilities" Dec 03 11:33:32 crc kubenswrapper[4646]: E1203 11:33:32.363769 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3fec42e-082c-406e-bf52-a6bf399c5c8b" containerName="registry-server" Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.363778 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3fec42e-082c-406e-bf52-a6bf399c5c8b" containerName="registry-server" Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.364022 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3fec42e-082c-406e-bf52-a6bf399c5c8b" containerName="registry-server" Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.365405 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-74vwp/must-gather-fxg6j" Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.368817 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-74vwp"/"kube-root-ca.crt" Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.376267 4646 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-74vwp"/"openshift-service-ca.crt" Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.393676 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-74vwp/must-gather-fxg6j"] Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.439253 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6da83524-87b7-401e-ade3-d99a262721e2-must-gather-output\") pod \"must-gather-fxg6j\" (UID: \"6da83524-87b7-401e-ade3-d99a262721e2\") " pod="openshift-must-gather-74vwp/must-gather-fxg6j" Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.439585 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkjv5\" (UniqueName: \"kubernetes.io/projected/6da83524-87b7-401e-ade3-d99a262721e2-kube-api-access-xkjv5\") pod \"must-gather-fxg6j\" (UID: \"6da83524-87b7-401e-ade3-d99a262721e2\") " pod="openshift-must-gather-74vwp/must-gather-fxg6j" Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.541937 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkjv5\" (UniqueName: \"kubernetes.io/projected/6da83524-87b7-401e-ade3-d99a262721e2-kube-api-access-xkjv5\") pod \"must-gather-fxg6j\" (UID: \"6da83524-87b7-401e-ade3-d99a262721e2\") " pod="openshift-must-gather-74vwp/must-gather-fxg6j" Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.542038 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6da83524-87b7-401e-ade3-d99a262721e2-must-gather-output\") pod \"must-gather-fxg6j\" (UID: \"6da83524-87b7-401e-ade3-d99a262721e2\") " pod="openshift-must-gather-74vwp/must-gather-fxg6j" Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.542585 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6da83524-87b7-401e-ade3-d99a262721e2-must-gather-output\") pod \"must-gather-fxg6j\" (UID: \"6da83524-87b7-401e-ade3-d99a262721e2\") " pod="openshift-must-gather-74vwp/must-gather-fxg6j" Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.564515 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkjv5\" (UniqueName: \"kubernetes.io/projected/6da83524-87b7-401e-ade3-d99a262721e2-kube-api-access-xkjv5\") pod \"must-gather-fxg6j\" (UID: \"6da83524-87b7-401e-ade3-d99a262721e2\") " pod="openshift-must-gather-74vwp/must-gather-fxg6j" Dec 03 11:33:32 crc kubenswrapper[4646]: I1203 11:33:32.682907 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-74vwp/must-gather-fxg6j" Dec 03 11:33:33 crc kubenswrapper[4646]: I1203 11:33:33.169642 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-74vwp/must-gather-fxg6j"] Dec 03 11:33:33 crc kubenswrapper[4646]: I1203 11:33:33.904040 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-74vwp/must-gather-fxg6j" event={"ID":"6da83524-87b7-401e-ade3-d99a262721e2","Type":"ContainerStarted","Data":"6119afe5f9a8d238dbfac32904a84e6d493276c49e5426015a677c16244461cb"} Dec 03 11:33:37 crc kubenswrapper[4646]: I1203 11:33:37.937544 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-74vwp/must-gather-fxg6j" event={"ID":"6da83524-87b7-401e-ade3-d99a262721e2","Type":"ContainerStarted","Data":"cbcc1dc6529ab7e7f3a79b081cb61dd2cf8f17cdac6a163d545dcfad7b7b0bc4"} Dec 03 11:33:38 crc kubenswrapper[4646]: I1203 11:33:38.948501 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-74vwp/must-gather-fxg6j" event={"ID":"6da83524-87b7-401e-ade3-d99a262721e2","Type":"ContainerStarted","Data":"0b94397d08edfaf51ebc7128a4185554039424a85dcbd6c2c709a77ab5e131b6"} Dec 03 11:33:38 crc kubenswrapper[4646]: I1203 11:33:38.972438 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-74vwp/must-gather-fxg6j" podStartSLOduration=2.649340253 podStartE2EDuration="6.972416952s" podCreationTimestamp="2025-12-03 11:33:32 +0000 UTC" firstStartedPulling="2025-12-03 11:33:33.188207499 +0000 UTC m=+2389.651263644" lastFinishedPulling="2025-12-03 11:33:37.511284208 +0000 UTC m=+2393.974340343" observedRunningTime="2025-12-03 11:33:38.969070907 +0000 UTC m=+2395.432127062" watchObservedRunningTime="2025-12-03 11:33:38.972416952 +0000 UTC m=+2395.435473087" Dec 03 11:33:41 crc kubenswrapper[4646]: I1203 11:33:41.590859 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-74vwp/crc-debug-bmrvf"] Dec 03 11:33:41 crc kubenswrapper[4646]: I1203 11:33:41.592951 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-74vwp/crc-debug-bmrvf" Dec 03 11:33:41 crc kubenswrapper[4646]: I1203 11:33:41.598699 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-74vwp"/"default-dockercfg-nfqr2" Dec 03 11:33:41 crc kubenswrapper[4646]: I1203 11:33:41.758554 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/205090dc-d026-4b1a-889b-51fa757b5136-host\") pod \"crc-debug-bmrvf\" (UID: \"205090dc-d026-4b1a-889b-51fa757b5136\") " pod="openshift-must-gather-74vwp/crc-debug-bmrvf" Dec 03 11:33:41 crc kubenswrapper[4646]: I1203 11:33:41.758951 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pttjt\" (UniqueName: \"kubernetes.io/projected/205090dc-d026-4b1a-889b-51fa757b5136-kube-api-access-pttjt\") pod \"crc-debug-bmrvf\" (UID: \"205090dc-d026-4b1a-889b-51fa757b5136\") " pod="openshift-must-gather-74vwp/crc-debug-bmrvf" Dec 03 11:33:41 crc kubenswrapper[4646]: I1203 11:33:41.860134 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pttjt\" (UniqueName: \"kubernetes.io/projected/205090dc-d026-4b1a-889b-51fa757b5136-kube-api-access-pttjt\") pod \"crc-debug-bmrvf\" (UID: \"205090dc-d026-4b1a-889b-51fa757b5136\") " pod="openshift-must-gather-74vwp/crc-debug-bmrvf" Dec 03 11:33:41 crc kubenswrapper[4646]: I1203 11:33:41.860289 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/205090dc-d026-4b1a-889b-51fa757b5136-host\") pod \"crc-debug-bmrvf\" (UID: \"205090dc-d026-4b1a-889b-51fa757b5136\") " pod="openshift-must-gather-74vwp/crc-debug-bmrvf" Dec 03 11:33:41 crc kubenswrapper[4646]: I1203 11:33:41.860444 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/205090dc-d026-4b1a-889b-51fa757b5136-host\") pod \"crc-debug-bmrvf\" (UID: \"205090dc-d026-4b1a-889b-51fa757b5136\") " pod="openshift-must-gather-74vwp/crc-debug-bmrvf" Dec 03 11:33:41 crc kubenswrapper[4646]: I1203 11:33:41.887127 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pttjt\" (UniqueName: \"kubernetes.io/projected/205090dc-d026-4b1a-889b-51fa757b5136-kube-api-access-pttjt\") pod \"crc-debug-bmrvf\" (UID: \"205090dc-d026-4b1a-889b-51fa757b5136\") " pod="openshift-must-gather-74vwp/crc-debug-bmrvf" Dec 03 11:33:41 crc kubenswrapper[4646]: I1203 11:33:41.910069 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-74vwp/crc-debug-bmrvf" Dec 03 11:33:41 crc kubenswrapper[4646]: W1203 11:33:41.953133 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod205090dc_d026_4b1a_889b_51fa757b5136.slice/crio-aecd9c583cc19378b7091959cb39bfae1c0e46a2aba4c57abe77a463cd61f91c WatchSource:0}: Error finding container aecd9c583cc19378b7091959cb39bfae1c0e46a2aba4c57abe77a463cd61f91c: Status 404 returned error can't find the container with id aecd9c583cc19378b7091959cb39bfae1c0e46a2aba4c57abe77a463cd61f91c Dec 03 11:33:41 crc kubenswrapper[4646]: I1203 11:33:41.972062 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-74vwp/crc-debug-bmrvf" event={"ID":"205090dc-d026-4b1a-889b-51fa757b5136","Type":"ContainerStarted","Data":"aecd9c583cc19378b7091959cb39bfae1c0e46a2aba4c57abe77a463cd61f91c"} Dec 03 11:33:54 crc kubenswrapper[4646]: I1203 11:33:54.095050 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-74vwp/crc-debug-bmrvf" event={"ID":"205090dc-d026-4b1a-889b-51fa757b5136","Type":"ContainerStarted","Data":"af927df07942013e99bc2b92694fc8f33b0fb759083a655c38a25e5e58e3e1be"} Dec 03 11:33:54 crc kubenswrapper[4646]: I1203 11:33:54.120457 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-74vwp/crc-debug-bmrvf" podStartSLOduration=1.274138711 podStartE2EDuration="13.120437824s" podCreationTimestamp="2025-12-03 11:33:41 +0000 UTC" firstStartedPulling="2025-12-03 11:33:41.95457931 +0000 UTC m=+2398.417635445" lastFinishedPulling="2025-12-03 11:33:53.800878423 +0000 UTC m=+2410.263934558" observedRunningTime="2025-12-03 11:33:54.111962812 +0000 UTC m=+2410.575018947" watchObservedRunningTime="2025-12-03 11:33:54.120437824 +0000 UTC m=+2410.583493959" Dec 03 11:33:55 crc kubenswrapper[4646]: I1203 11:33:55.964593 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:33:55 crc kubenswrapper[4646]: I1203 11:33:55.965146 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:33:55 crc kubenswrapper[4646]: I1203 11:33:55.965190 4646 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 11:33:55 crc kubenswrapper[4646]: I1203 11:33:55.965918 4646 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29"} pod="openshift-machine-config-operator/machine-config-daemon-pndzt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 11:33:55 crc kubenswrapper[4646]: I1203 11:33:55.965965 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" containerID="cri-o://19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" gracePeriod=600 Dec 03 11:33:56 crc kubenswrapper[4646]: E1203 11:33:56.098377 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:33:56 crc kubenswrapper[4646]: I1203 11:33:56.121647 4646 generic.go:334] "Generic (PLEG): container finished" podID="9695d947-fc43-4625-86af-a1a3e177bb27" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" exitCode=0 Dec 03 11:33:56 crc kubenswrapper[4646]: I1203 11:33:56.121692 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerDied","Data":"19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29"} Dec 03 11:33:56 crc kubenswrapper[4646]: I1203 11:33:56.121728 4646 scope.go:117] "RemoveContainer" containerID="52c3eac8f06b635ff3f5d9230f530cf2d7c87e4c968e13300ddfce7f0bba1dcb" Dec 03 11:33:56 crc kubenswrapper[4646]: I1203 11:33:56.122253 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:33:56 crc kubenswrapper[4646]: E1203 11:33:56.122534 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:34:06 crc kubenswrapper[4646]: I1203 11:34:06.849225 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:34:06 crc kubenswrapper[4646]: E1203 11:34:06.849879 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:34:10 crc kubenswrapper[4646]: I1203 11:34:10.234225 4646 generic.go:334] "Generic (PLEG): container finished" podID="205090dc-d026-4b1a-889b-51fa757b5136" containerID="af927df07942013e99bc2b92694fc8f33b0fb759083a655c38a25e5e58e3e1be" exitCode=0 Dec 03 11:34:10 crc kubenswrapper[4646]: I1203 11:34:10.234322 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-74vwp/crc-debug-bmrvf" event={"ID":"205090dc-d026-4b1a-889b-51fa757b5136","Type":"ContainerDied","Data":"af927df07942013e99bc2b92694fc8f33b0fb759083a655c38a25e5e58e3e1be"} Dec 03 11:34:11 crc kubenswrapper[4646]: I1203 11:34:11.357927 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-74vwp/crc-debug-bmrvf" Dec 03 11:34:11 crc kubenswrapper[4646]: I1203 11:34:11.395234 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-74vwp/crc-debug-bmrvf"] Dec 03 11:34:11 crc kubenswrapper[4646]: I1203 11:34:11.429192 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-74vwp/crc-debug-bmrvf"] Dec 03 11:34:11 crc kubenswrapper[4646]: I1203 11:34:11.491392 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/205090dc-d026-4b1a-889b-51fa757b5136-host\") pod \"205090dc-d026-4b1a-889b-51fa757b5136\" (UID: \"205090dc-d026-4b1a-889b-51fa757b5136\") " Dec 03 11:34:11 crc kubenswrapper[4646]: I1203 11:34:11.491558 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pttjt\" (UniqueName: \"kubernetes.io/projected/205090dc-d026-4b1a-889b-51fa757b5136-kube-api-access-pttjt\") pod \"205090dc-d026-4b1a-889b-51fa757b5136\" (UID: \"205090dc-d026-4b1a-889b-51fa757b5136\") " Dec 03 11:34:11 crc kubenswrapper[4646]: I1203 11:34:11.491777 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/205090dc-d026-4b1a-889b-51fa757b5136-host" (OuterVolumeSpecName: "host") pod "205090dc-d026-4b1a-889b-51fa757b5136" (UID: "205090dc-d026-4b1a-889b-51fa757b5136"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:34:11 crc kubenswrapper[4646]: I1203 11:34:11.492264 4646 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/205090dc-d026-4b1a-889b-51fa757b5136-host\") on node \"crc\" DevicePath \"\"" Dec 03 11:34:11 crc kubenswrapper[4646]: I1203 11:34:11.504173 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/205090dc-d026-4b1a-889b-51fa757b5136-kube-api-access-pttjt" (OuterVolumeSpecName: "kube-api-access-pttjt") pod "205090dc-d026-4b1a-889b-51fa757b5136" (UID: "205090dc-d026-4b1a-889b-51fa757b5136"). InnerVolumeSpecName "kube-api-access-pttjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:34:11 crc kubenswrapper[4646]: I1203 11:34:11.593843 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pttjt\" (UniqueName: \"kubernetes.io/projected/205090dc-d026-4b1a-889b-51fa757b5136-kube-api-access-pttjt\") on node \"crc\" DevicePath \"\"" Dec 03 11:34:11 crc kubenswrapper[4646]: I1203 11:34:11.860008 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="205090dc-d026-4b1a-889b-51fa757b5136" path="/var/lib/kubelet/pods/205090dc-d026-4b1a-889b-51fa757b5136/volumes" Dec 03 11:34:12 crc kubenswrapper[4646]: I1203 11:34:12.252444 4646 scope.go:117] "RemoveContainer" containerID="af927df07942013e99bc2b92694fc8f33b0fb759083a655c38a25e5e58e3e1be" Dec 03 11:34:12 crc kubenswrapper[4646]: I1203 11:34:12.252505 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-74vwp/crc-debug-bmrvf" Dec 03 11:34:12 crc kubenswrapper[4646]: I1203 11:34:12.632434 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-74vwp/crc-debug-4ch5z"] Dec 03 11:34:12 crc kubenswrapper[4646]: E1203 11:34:12.632805 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="205090dc-d026-4b1a-889b-51fa757b5136" containerName="container-00" Dec 03 11:34:12 crc kubenswrapper[4646]: I1203 11:34:12.632816 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="205090dc-d026-4b1a-889b-51fa757b5136" containerName="container-00" Dec 03 11:34:12 crc kubenswrapper[4646]: I1203 11:34:12.632961 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="205090dc-d026-4b1a-889b-51fa757b5136" containerName="container-00" Dec 03 11:34:12 crc kubenswrapper[4646]: I1203 11:34:12.633535 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-74vwp/crc-debug-4ch5z" Dec 03 11:34:12 crc kubenswrapper[4646]: I1203 11:34:12.637905 4646 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-74vwp"/"default-dockercfg-nfqr2" Dec 03 11:34:12 crc kubenswrapper[4646]: I1203 11:34:12.813953 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/086eff50-3a10-4de4-8fd7-3ecfd54e5a25-host\") pod \"crc-debug-4ch5z\" (UID: \"086eff50-3a10-4de4-8fd7-3ecfd54e5a25\") " pod="openshift-must-gather-74vwp/crc-debug-4ch5z" Dec 03 11:34:12 crc kubenswrapper[4646]: I1203 11:34:12.814042 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46pwl\" (UniqueName: \"kubernetes.io/projected/086eff50-3a10-4de4-8fd7-3ecfd54e5a25-kube-api-access-46pwl\") pod \"crc-debug-4ch5z\" (UID: \"086eff50-3a10-4de4-8fd7-3ecfd54e5a25\") " pod="openshift-must-gather-74vwp/crc-debug-4ch5z" Dec 03 11:34:12 crc kubenswrapper[4646]: I1203 11:34:12.916107 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46pwl\" (UniqueName: \"kubernetes.io/projected/086eff50-3a10-4de4-8fd7-3ecfd54e5a25-kube-api-access-46pwl\") pod \"crc-debug-4ch5z\" (UID: \"086eff50-3a10-4de4-8fd7-3ecfd54e5a25\") " pod="openshift-must-gather-74vwp/crc-debug-4ch5z" Dec 03 11:34:12 crc kubenswrapper[4646]: I1203 11:34:12.916368 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/086eff50-3a10-4de4-8fd7-3ecfd54e5a25-host\") pod \"crc-debug-4ch5z\" (UID: \"086eff50-3a10-4de4-8fd7-3ecfd54e5a25\") " pod="openshift-must-gather-74vwp/crc-debug-4ch5z" Dec 03 11:34:12 crc kubenswrapper[4646]: I1203 11:34:12.916499 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/086eff50-3a10-4de4-8fd7-3ecfd54e5a25-host\") pod \"crc-debug-4ch5z\" (UID: \"086eff50-3a10-4de4-8fd7-3ecfd54e5a25\") " pod="openshift-must-gather-74vwp/crc-debug-4ch5z" Dec 03 11:34:12 crc kubenswrapper[4646]: I1203 11:34:12.934366 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46pwl\" (UniqueName: \"kubernetes.io/projected/086eff50-3a10-4de4-8fd7-3ecfd54e5a25-kube-api-access-46pwl\") pod \"crc-debug-4ch5z\" (UID: \"086eff50-3a10-4de4-8fd7-3ecfd54e5a25\") " pod="openshift-must-gather-74vwp/crc-debug-4ch5z" Dec 03 11:34:12 crc kubenswrapper[4646]: I1203 11:34:12.948348 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-74vwp/crc-debug-4ch5z" Dec 03 11:34:13 crc kubenswrapper[4646]: I1203 11:34:13.262608 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-74vwp/crc-debug-4ch5z" event={"ID":"086eff50-3a10-4de4-8fd7-3ecfd54e5a25","Type":"ContainerStarted","Data":"1f907344f72a2a4d3651183bbf22f3ecada000183c9314c325213dccd5487f8c"} Dec 03 11:34:14 crc kubenswrapper[4646]: I1203 11:34:14.273956 4646 generic.go:334] "Generic (PLEG): container finished" podID="086eff50-3a10-4de4-8fd7-3ecfd54e5a25" containerID="219b7acfd1a0af8ab685e6056652ba8396d9101018e133b2aadafbecd077e6ad" exitCode=1 Dec 03 11:34:14 crc kubenswrapper[4646]: I1203 11:34:14.274014 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-74vwp/crc-debug-4ch5z" event={"ID":"086eff50-3a10-4de4-8fd7-3ecfd54e5a25","Type":"ContainerDied","Data":"219b7acfd1a0af8ab685e6056652ba8396d9101018e133b2aadafbecd077e6ad"} Dec 03 11:34:14 crc kubenswrapper[4646]: I1203 11:34:14.317698 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-74vwp/crc-debug-4ch5z"] Dec 03 11:34:14 crc kubenswrapper[4646]: I1203 11:34:14.326111 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-74vwp/crc-debug-4ch5z"] Dec 03 11:34:15 crc kubenswrapper[4646]: I1203 11:34:15.409688 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-74vwp/crc-debug-4ch5z" Dec 03 11:34:15 crc kubenswrapper[4646]: I1203 11:34:15.570270 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46pwl\" (UniqueName: \"kubernetes.io/projected/086eff50-3a10-4de4-8fd7-3ecfd54e5a25-kube-api-access-46pwl\") pod \"086eff50-3a10-4de4-8fd7-3ecfd54e5a25\" (UID: \"086eff50-3a10-4de4-8fd7-3ecfd54e5a25\") " Dec 03 11:34:15 crc kubenswrapper[4646]: I1203 11:34:15.570771 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/086eff50-3a10-4de4-8fd7-3ecfd54e5a25-host\") pod \"086eff50-3a10-4de4-8fd7-3ecfd54e5a25\" (UID: \"086eff50-3a10-4de4-8fd7-3ecfd54e5a25\") " Dec 03 11:34:15 crc kubenswrapper[4646]: I1203 11:34:15.570942 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/086eff50-3a10-4de4-8fd7-3ecfd54e5a25-host" (OuterVolumeSpecName: "host") pod "086eff50-3a10-4de4-8fd7-3ecfd54e5a25" (UID: "086eff50-3a10-4de4-8fd7-3ecfd54e5a25"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 11:34:15 crc kubenswrapper[4646]: I1203 11:34:15.571392 4646 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/086eff50-3a10-4de4-8fd7-3ecfd54e5a25-host\") on node \"crc\" DevicePath \"\"" Dec 03 11:34:15 crc kubenswrapper[4646]: I1203 11:34:15.580567 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/086eff50-3a10-4de4-8fd7-3ecfd54e5a25-kube-api-access-46pwl" (OuterVolumeSpecName: "kube-api-access-46pwl") pod "086eff50-3a10-4de4-8fd7-3ecfd54e5a25" (UID: "086eff50-3a10-4de4-8fd7-3ecfd54e5a25"). InnerVolumeSpecName "kube-api-access-46pwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:34:15 crc kubenswrapper[4646]: I1203 11:34:15.673611 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46pwl\" (UniqueName: \"kubernetes.io/projected/086eff50-3a10-4de4-8fd7-3ecfd54e5a25-kube-api-access-46pwl\") on node \"crc\" DevicePath \"\"" Dec 03 11:34:15 crc kubenswrapper[4646]: I1203 11:34:15.858036 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="086eff50-3a10-4de4-8fd7-3ecfd54e5a25" path="/var/lib/kubelet/pods/086eff50-3a10-4de4-8fd7-3ecfd54e5a25/volumes" Dec 03 11:34:16 crc kubenswrapper[4646]: I1203 11:34:16.290479 4646 scope.go:117] "RemoveContainer" containerID="219b7acfd1a0af8ab685e6056652ba8396d9101018e133b2aadafbecd077e6ad" Dec 03 11:34:16 crc kubenswrapper[4646]: I1203 11:34:16.290537 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-74vwp/crc-debug-4ch5z" Dec 03 11:34:17 crc kubenswrapper[4646]: I1203 11:34:17.848593 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:34:17 crc kubenswrapper[4646]: E1203 11:34:17.849319 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:34:30 crc kubenswrapper[4646]: I1203 11:34:30.848662 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:34:30 crc kubenswrapper[4646]: E1203 11:34:30.849394 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:34:43 crc kubenswrapper[4646]: I1203 11:34:43.885528 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:34:43 crc kubenswrapper[4646]: E1203 11:34:43.902823 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:34:55 crc kubenswrapper[4646]: I1203 11:34:55.849842 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:34:55 crc kubenswrapper[4646]: E1203 11:34:55.850594 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:35:02 crc kubenswrapper[4646]: I1203 11:35:02.245994 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6d775b8f86-64hmf_07721619-ea6b-4cb8-95bb-f5801fc848e3/barbican-api/0.log" Dec 03 11:35:02 crc kubenswrapper[4646]: I1203 11:35:02.249654 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6d775b8f86-64hmf_07721619-ea6b-4cb8-95bb-f5801fc848e3/barbican-api-log/0.log" Dec 03 11:35:02 crc kubenswrapper[4646]: I1203 11:35:02.551486 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b6bfd6548-5hldq_10ed882e-df14-4619-9a57-022edbefa36c/barbican-keystone-listener-log/0.log" Dec 03 11:35:02 crc kubenswrapper[4646]: I1203 11:35:02.568292 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b6bfd6548-5hldq_10ed882e-df14-4619-9a57-022edbefa36c/barbican-keystone-listener/0.log" Dec 03 11:35:02 crc kubenswrapper[4646]: I1203 11:35:02.723160 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9f949dfbc-7rr98_eaba04e4-bc4a-41a0-aa79-631669451754/barbican-worker/0.log" Dec 03 11:35:02 crc kubenswrapper[4646]: I1203 11:35:02.894005 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9f949dfbc-7rr98_eaba04e4-bc4a-41a0-aa79-631669451754/barbican-worker-log/0.log" Dec 03 11:35:03 crc kubenswrapper[4646]: I1203 11:35:03.010632 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-bhdj5_d6fdb595-c620-4fd6-9a04-5acc542f158d/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 11:35:03 crc kubenswrapper[4646]: I1203 11:35:03.164547 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_317f0cc8-e513-472a-86c4-9f49b831bbba/ceilometer-central-agent/0.log" Dec 03 11:35:03 crc kubenswrapper[4646]: I1203 11:35:03.227732 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_317f0cc8-e513-472a-86c4-9f49b831bbba/ceilometer-notification-agent/0.log" Dec 03 11:35:03 crc kubenswrapper[4646]: I1203 11:35:03.283392 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_317f0cc8-e513-472a-86c4-9f49b831bbba/sg-core/0.log" Dec 03 11:35:03 crc kubenswrapper[4646]: I1203 11:35:03.326227 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_317f0cc8-e513-472a-86c4-9f49b831bbba/proxy-httpd/0.log" Dec 03 11:35:03 crc kubenswrapper[4646]: I1203 11:35:03.508863 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-djn64_246b95ce-fd50-4018-b4f1-329c2c58d1fc/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 11:35:03 crc kubenswrapper[4646]: I1203 11:35:03.639247 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ad772e53-d750-4f6c-99be-ff551ebd0669/cinder-api/0.log" Dec 03 11:35:03 crc kubenswrapper[4646]: I1203 11:35:03.790069 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ad772e53-d750-4f6c-99be-ff551ebd0669/cinder-api-log/0.log" Dec 03 11:35:03 crc kubenswrapper[4646]: I1203 11:35:03.866451 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e47516a7-a5ef-44f9-883e-b8c0383851c1/probe/0.log" Dec 03 11:35:03 crc kubenswrapper[4646]: I1203 11:35:03.965452 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e47516a7-a5ef-44f9-883e-b8c0383851c1/cinder-scheduler/0.log" Dec 03 11:35:04 crc kubenswrapper[4646]: I1203 11:35:04.123572 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-hgcdm_6185a07d-c02f-42b5-b060-c5317850e67b/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 11:35:04 crc kubenswrapper[4646]: I1203 11:35:04.229263 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-hxhwq_9320c819-432d-4440-8897-659b2fa98a20/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 11:35:04 crc kubenswrapper[4646]: I1203 11:35:04.378019 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-69fd9b48bc-x88c4_b4c02b37-8c88-41d6-866f-e31ef5cdfab3/init/0.log" Dec 03 11:35:04 crc kubenswrapper[4646]: I1203 11:35:04.648002 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-69fd9b48bc-x88c4_b4c02b37-8c88-41d6-866f-e31ef5cdfab3/init/0.log" Dec 03 11:35:04 crc kubenswrapper[4646]: I1203 11:35:04.695088 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-69fd9b48bc-x88c4_b4c02b37-8c88-41d6-866f-e31ef5cdfab3/dnsmasq-dns/0.log" Dec 03 11:35:04 crc kubenswrapper[4646]: I1203 11:35:04.799748 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-dn9j6_66d26829-4e09-477d-920a-07908ee19a0a/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 11:35:04 crc kubenswrapper[4646]: I1203 11:35:04.974124 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-779dc4bcb9-gdhmd_7f051542-4055-4cde-aaf3-8c30075cb832/keystone-api/0.log" Dec 03 11:35:05 crc kubenswrapper[4646]: I1203 11:35:05.348066 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_eb37f32e-cad3-4184-ad71-fd5cef58ca1f/kube-state-metrics/0.log" Dec 03 11:35:05 crc kubenswrapper[4646]: I1203 11:35:05.617191 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-59d75d495c-x9phz_c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157/neutron-api/0.log" Dec 03 11:35:06 crc kubenswrapper[4646]: I1203 11:35:06.067355 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-59d75d495c-x9phz_c6abb2d5-b3a1-46bf-a9fd-b0fa9cb15157/neutron-httpd/0.log" Dec 03 11:35:06 crc kubenswrapper[4646]: I1203 11:35:06.322533 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_99c9b16d-acd9-453f-adb9-505992d25ec2/nova-api-api/0.log" Dec 03 11:35:06 crc kubenswrapper[4646]: I1203 11:35:06.351888 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_99c9b16d-acd9-453f-adb9-505992d25ec2/nova-api-log/0.log" Dec 03 11:35:06 crc kubenswrapper[4646]: I1203 11:35:06.698762 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_c0b8e9c3-6587-4b17-a294-9d923deb1047/nova-cell0-conductor-conductor/0.log" Dec 03 11:35:06 crc kubenswrapper[4646]: I1203 11:35:06.823107 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_82149567-e81e-4056-829c-a854bee435c5/nova-cell1-conductor-conductor/0.log" Dec 03 11:35:07 crc kubenswrapper[4646]: I1203 11:35:07.046268 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9a064e62-8fce-4cb7-9d19-e1e0542570c3/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 11:35:07 crc kubenswrapper[4646]: I1203 11:35:07.290407 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_27ad9778-57c8-4308-95e8-c20592f589c7/nova-metadata-log/0.log" Dec 03 11:35:07 crc kubenswrapper[4646]: I1203 11:35:07.833457 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_f7bfc0c4-44a4-4f09-8a04-22fad18a1e69/nova-scheduler-scheduler/0.log" Dec 03 11:35:07 crc kubenswrapper[4646]: I1203 11:35:07.870551 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0497fe5e-a52c-4f64-9517-e10eb809829a/mysql-bootstrap/0.log" Dec 03 11:35:07 crc kubenswrapper[4646]: I1203 11:35:07.993054 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_27ad9778-57c8-4308-95e8-c20592f589c7/nova-metadata-metadata/0.log" Dec 03 11:35:08 crc kubenswrapper[4646]: I1203 11:35:08.192647 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0497fe5e-a52c-4f64-9517-e10eb809829a/mysql-bootstrap/0.log" Dec 03 11:35:08 crc kubenswrapper[4646]: I1203 11:35:08.233314 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0497fe5e-a52c-4f64-9517-e10eb809829a/galera/0.log" Dec 03 11:35:08 crc kubenswrapper[4646]: I1203 11:35:08.356950 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d29b2323-9322-47f0-983c-bba7a541f4e5/mysql-bootstrap/0.log" Dec 03 11:35:08 crc kubenswrapper[4646]: I1203 11:35:08.535428 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d29b2323-9322-47f0-983c-bba7a541f4e5/mysql-bootstrap/0.log" Dec 03 11:35:08 crc kubenswrapper[4646]: I1203 11:35:08.605773 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d29b2323-9322-47f0-983c-bba7a541f4e5/galera/0.log" Dec 03 11:35:08 crc kubenswrapper[4646]: I1203 11:35:08.693563 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_4c18061a-8b6e-4c46-9475-2a2fbba4bd7a/openstackclient/0.log" Dec 03 11:35:08 crc kubenswrapper[4646]: I1203 11:35:08.997646 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-lg5vm_b03f673e-7c4b-4d20-b297-be13a71f8e42/openstack-network-exporter/0.log" Dec 03 11:35:08 crc kubenswrapper[4646]: I1203 11:35:08.999526 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5v9pv_1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4/ovsdb-server-init/0.log" Dec 03 11:35:09 crc kubenswrapper[4646]: I1203 11:35:09.619394 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5v9pv_1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4/ovs-vswitchd/0.log" Dec 03 11:35:09 crc kubenswrapper[4646]: I1203 11:35:09.632424 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5v9pv_1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4/ovsdb-server-init/0.log" Dec 03 11:35:09 crc kubenswrapper[4646]: I1203 11:35:09.720718 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5v9pv_1e5b9d38-1bc8-4d85-9b9b-33b9ea0dadb4/ovsdb-server/0.log" Dec 03 11:35:09 crc kubenswrapper[4646]: I1203 11:35:09.955596 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-rjmpb_f7606fd8-32a7-402e-bac9-bfe718af9019/ovn-controller/0.log" Dec 03 11:35:10 crc kubenswrapper[4646]: I1203 11:35:10.019890 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_46f05884-7ce0-49e9-bf34-fdb200c78095/openstack-network-exporter/0.log" Dec 03 11:35:10 crc kubenswrapper[4646]: I1203 11:35:10.020757 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_46f05884-7ce0-49e9-bf34-fdb200c78095/ovn-northd/0.log" Dec 03 11:35:10 crc kubenswrapper[4646]: I1203 11:35:10.293199 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2ab2f042-5e66-4aa8-9a86-6687c40a675d/openstack-network-exporter/0.log" Dec 03 11:35:10 crc kubenswrapper[4646]: I1203 11:35:10.350708 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2ab2f042-5e66-4aa8-9a86-6687c40a675d/ovsdbserver-nb/0.log" Dec 03 11:35:10 crc kubenswrapper[4646]: I1203 11:35:10.570410 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6/ovsdbserver-sb/0.log" Dec 03 11:35:10 crc kubenswrapper[4646]: I1203 11:35:10.616438 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_155aca02-6cc2-40d2-9d5f-e29a6bc9c1f6/openstack-network-exporter/0.log" Dec 03 11:35:10 crc kubenswrapper[4646]: I1203 11:35:10.712910 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-96574894d-gjvg2_b862ee56-bca1-4ea4-b2e8-d1ef47365550/placement-api/0.log" Dec 03 11:35:10 crc kubenswrapper[4646]: I1203 11:35:10.848433 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:35:10 crc kubenswrapper[4646]: E1203 11:35:10.848632 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:35:10 crc kubenswrapper[4646]: I1203 11:35:10.955788 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_16b4b1d1-a293-4042-b52b-dc8a10e07536/setup-container/0.log" Dec 03 11:35:11 crc kubenswrapper[4646]: I1203 11:35:11.066915 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-96574894d-gjvg2_b862ee56-bca1-4ea4-b2e8-d1ef47365550/placement-log/0.log" Dec 03 11:35:11 crc kubenswrapper[4646]: I1203 11:35:11.298784 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_16b4b1d1-a293-4042-b52b-dc8a10e07536/rabbitmq/0.log" Dec 03 11:35:11 crc kubenswrapper[4646]: I1203 11:35:11.335817 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_16b4b1d1-a293-4042-b52b-dc8a10e07536/setup-container/0.log" Dec 03 11:35:11 crc kubenswrapper[4646]: I1203 11:35:11.353718 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3e83310b-1692-4be3-83ba-8278d0dc31e2/setup-container/0.log" Dec 03 11:35:11 crc kubenswrapper[4646]: I1203 11:35:11.684769 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3e83310b-1692-4be3-83ba-8278d0dc31e2/rabbitmq/0.log" Dec 03 11:35:11 crc kubenswrapper[4646]: I1203 11:35:11.738467 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3e83310b-1692-4be3-83ba-8278d0dc31e2/setup-container/0.log" Dec 03 11:35:11 crc kubenswrapper[4646]: I1203 11:35:11.765763 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-cvww9_a6724a5b-6c06-48df-a742-d9c0484ab526/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 11:35:11 crc kubenswrapper[4646]: I1203 11:35:11.995158 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-9rf6w_5afc0cac-1a45-4f71-94e6-241231108859/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 11:35:12 crc kubenswrapper[4646]: I1203 11:35:12.202651 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-9vsdg_f01298f9-3c81-4297-921f-0a201d0ac557/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 11:35:12 crc kubenswrapper[4646]: I1203 11:35:12.338933 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-7vhz8_5e44d032-6062-45a5-b3cd-59e54618238f/ssh-known-hosts-edpm-deployment/0.log" Dec 03 11:35:12 crc kubenswrapper[4646]: I1203 11:35:12.556123 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-wxl7n_9a1511a5-3474-484d-a936-3ed46c8b3ed2/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 11:35:12 crc kubenswrapper[4646]: I1203 11:35:12.614634 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_493009da-44cc-43ec-8164-4c1c20ab94e1/memcached/0.log" Dec 03 11:35:22 crc kubenswrapper[4646]: I1203 11:35:22.848156 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:35:22 crc kubenswrapper[4646]: E1203 11:35:22.849064 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:35:35 crc kubenswrapper[4646]: I1203 11:35:35.882318 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:35:35 crc kubenswrapper[4646]: E1203 11:35:35.883427 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:35:36 crc kubenswrapper[4646]: I1203 11:35:36.837041 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs_f9512558-d306-43f1-b530-6c25875a7625/util/0.log" Dec 03 11:35:37 crc kubenswrapper[4646]: I1203 11:35:37.070419 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs_f9512558-d306-43f1-b530-6c25875a7625/pull/0.log" Dec 03 11:35:37 crc kubenswrapper[4646]: I1203 11:35:37.093409 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs_f9512558-d306-43f1-b530-6c25875a7625/util/0.log" Dec 03 11:35:37 crc kubenswrapper[4646]: I1203 11:35:37.169965 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs_f9512558-d306-43f1-b530-6c25875a7625/pull/0.log" Dec 03 11:35:37 crc kubenswrapper[4646]: I1203 11:35:37.338992 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs_f9512558-d306-43f1-b530-6c25875a7625/extract/0.log" Dec 03 11:35:37 crc kubenswrapper[4646]: I1203 11:35:37.357430 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs_f9512558-d306-43f1-b530-6c25875a7625/pull/0.log" Dec 03 11:35:37 crc kubenswrapper[4646]: I1203 11:35:37.390810 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_93478431aa2b3d3b80c27233e6c01b4e808bbc7d94ec8fef50d094ae1ch7xjs_f9512558-d306-43f1-b530-6c25875a7625/util/0.log" Dec 03 11:35:37 crc kubenswrapper[4646]: I1203 11:35:37.623289 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-ss5tr_e400aa20-0649-47de-8f10-da6090800eb3/manager/0.log" Dec 03 11:35:37 crc kubenswrapper[4646]: I1203 11:35:37.666578 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-pnbk8_d7ed2042-fe69-40f6-8440-49d63e851ffc/kube-rbac-proxy/0.log" Dec 03 11:35:37 crc kubenswrapper[4646]: I1203 11:35:37.687408 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-ss5tr_e400aa20-0649-47de-8f10-da6090800eb3/kube-rbac-proxy/0.log" Dec 03 11:35:37 crc kubenswrapper[4646]: I1203 11:35:37.881602 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-pnbk8_d7ed2042-fe69-40f6-8440-49d63e851ffc/manager/0.log" Dec 03 11:35:37 crc kubenswrapper[4646]: I1203 11:35:37.959327 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-ps2zm_4cb5b92a-540a-41d9-a662-b1f3faf07829/kube-rbac-proxy/0.log" Dec 03 11:35:38 crc kubenswrapper[4646]: I1203 11:35:38.009833 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-ps2zm_4cb5b92a-540a-41d9-a662-b1f3faf07829/manager/0.log" Dec 03 11:35:38 crc kubenswrapper[4646]: I1203 11:35:38.114731 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-56r4n_1fed03a1-2388-470c-843e-4de04fd6d9bc/kube-rbac-proxy/0.log" Dec 03 11:35:38 crc kubenswrapper[4646]: I1203 11:35:38.231457 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-56r4n_1fed03a1-2388-470c-843e-4de04fd6d9bc/manager/0.log" Dec 03 11:35:38 crc kubenswrapper[4646]: I1203 11:35:38.347041 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-mzlrq_396bd18c-37f7-471f-a0f9-0b9cb5a447e3/kube-rbac-proxy/0.log" Dec 03 11:35:38 crc kubenswrapper[4646]: I1203 11:35:38.433517 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-mzlrq_396bd18c-37f7-471f-a0f9-0b9cb5a447e3/manager/0.log" Dec 03 11:35:38 crc kubenswrapper[4646]: I1203 11:35:38.547576 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-6hl22_089f6f35-ab06-4b31-a331-16411a9783e7/kube-rbac-proxy/0.log" Dec 03 11:35:38 crc kubenswrapper[4646]: I1203 11:35:38.575905 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-6hl22_089f6f35-ab06-4b31-a331-16411a9783e7/manager/0.log" Dec 03 11:35:38 crc kubenswrapper[4646]: I1203 11:35:38.744350 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-g6jbp_d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0/kube-rbac-proxy/0.log" Dec 03 11:35:38 crc kubenswrapper[4646]: I1203 11:35:38.959856 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-g6jbp_d5dfec9c-60de-45c2-9e93-d3f9ef6fabb0/manager/0.log" Dec 03 11:35:39 crc kubenswrapper[4646]: I1203 11:35:39.026072 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-schbl_31371552-226a-436f-a595-52bc387ed5cf/kube-rbac-proxy/0.log" Dec 03 11:35:39 crc kubenswrapper[4646]: I1203 11:35:39.053601 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-schbl_31371552-226a-436f-a595-52bc387ed5cf/manager/0.log" Dec 03 11:35:39 crc kubenswrapper[4646]: I1203 11:35:39.189146 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-l7v6k_68e6047b-627f-4c45-aa55-4aeab90e17b5/kube-rbac-proxy/0.log" Dec 03 11:35:39 crc kubenswrapper[4646]: I1203 11:35:39.299067 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-l7v6k_68e6047b-627f-4c45-aa55-4aeab90e17b5/manager/0.log" Dec 03 11:35:39 crc kubenswrapper[4646]: I1203 11:35:39.644980 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-rcmlq_f3e07f77-c87f-4c6b-be4e-938d71354f8b/kube-rbac-proxy/0.log" Dec 03 11:35:39 crc kubenswrapper[4646]: I1203 11:35:39.737017 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-92vgw_a0fa06f6-534e-4ad4-806d-c13cad106809/kube-rbac-proxy/0.log" Dec 03 11:35:39 crc kubenswrapper[4646]: I1203 11:35:39.737639 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-rcmlq_f3e07f77-c87f-4c6b-be4e-938d71354f8b/manager/0.log" Dec 03 11:35:39 crc kubenswrapper[4646]: I1203 11:35:39.991802 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-92vgw_a0fa06f6-534e-4ad4-806d-c13cad106809/manager/0.log" Dec 03 11:35:40 crc kubenswrapper[4646]: I1203 11:35:40.014993 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-sc2vn_eaefadf5-75cd-4ff2-811c-e25bb2850173/kube-rbac-proxy/0.log" Dec 03 11:35:40 crc kubenswrapper[4646]: I1203 11:35:40.064348 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-sc2vn_eaefadf5-75cd-4ff2-811c-e25bb2850173/manager/0.log" Dec 03 11:35:40 crc kubenswrapper[4646]: I1203 11:35:40.335392 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-g8pg8_9a1416d1-243f-4ab9-9a9c-4b71e0f23876/manager/0.log" Dec 03 11:35:40 crc kubenswrapper[4646]: I1203 11:35:40.376241 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-g8pg8_9a1416d1-243f-4ab9-9a9c-4b71e0f23876/kube-rbac-proxy/0.log" Dec 03 11:35:40 crc kubenswrapper[4646]: I1203 11:35:40.532935 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-n9s99_ed761af7-289e-4f6c-a7b9-710886250a3a/kube-rbac-proxy/0.log" Dec 03 11:35:40 crc kubenswrapper[4646]: I1203 11:35:40.579403 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-n9s99_ed761af7-289e-4f6c-a7b9-710886250a3a/manager/0.log" Dec 03 11:35:40 crc kubenswrapper[4646]: I1203 11:35:40.676157 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9_e776d6da-584d-498f-b611-0eb67be6d6d0/kube-rbac-proxy/0.log" Dec 03 11:35:40 crc kubenswrapper[4646]: I1203 11:35:40.825245 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4cb4w9_e776d6da-584d-498f-b611-0eb67be6d6d0/manager/0.log" Dec 03 11:35:41 crc kubenswrapper[4646]: I1203 11:35:41.292937 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zdrkq_a15c5248-9b54-4744-b8e9-f411b6a3c1a1/registry-server/0.log" Dec 03 11:35:41 crc kubenswrapper[4646]: I1203 11:35:41.356633 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-779dc79ddf-9gqzp_509a9181-c788-464c-b157-4928132fadbe/operator/0.log" Dec 03 11:35:41 crc kubenswrapper[4646]: I1203 11:35:41.500304 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-t7fj8_153f2bfa-6ca2-4a9f-900e-5321c5791639/kube-rbac-proxy/0.log" Dec 03 11:35:41 crc kubenswrapper[4646]: I1203 11:35:41.577155 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-t7fj8_153f2bfa-6ca2-4a9f-900e-5321c5791639/manager/0.log" Dec 03 11:35:41 crc kubenswrapper[4646]: I1203 11:35:41.590101 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-f7pcq_011cd55c-5e76-4dd3-905b-62c587119096/kube-rbac-proxy/0.log" Dec 03 11:35:41 crc kubenswrapper[4646]: I1203 11:35:41.728293 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6c5c989645-kkkb7_83f119ed-6576-4e47-82b1-a06189cff0a7/manager/0.log" Dec 03 11:35:41 crc kubenswrapper[4646]: I1203 11:35:41.814028 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-f7pcq_011cd55c-5e76-4dd3-905b-62c587119096/manager/0.log" Dec 03 11:35:41 crc kubenswrapper[4646]: I1203 11:35:41.893862 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-b5vdh_afd2b4e9-59a4-4e52-bd72-706b8edbd719/operator/0.log" Dec 03 11:35:41 crc kubenswrapper[4646]: I1203 11:35:41.995626 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-j2bwb_f274e7e0-a406-4617-95bc-fbca7ef39526/kube-rbac-proxy/0.log" Dec 03 11:35:42 crc kubenswrapper[4646]: I1203 11:35:42.024282 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-j2bwb_f274e7e0-a406-4617-95bc-fbca7ef39526/manager/0.log" Dec 03 11:35:42 crc kubenswrapper[4646]: I1203 11:35:42.079247 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-w5c45_55e5ed61-839c-441b-acf2-cd43cc0c5c90/kube-rbac-proxy/0.log" Dec 03 11:35:42 crc kubenswrapper[4646]: I1203 11:35:42.216535 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-w5c45_55e5ed61-839c-441b-acf2-cd43cc0c5c90/manager/0.log" Dec 03 11:35:42 crc kubenswrapper[4646]: I1203 11:35:42.246897 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-z6m86_0be433f8-aae1-48a4-8853-de55ae48f836/kube-rbac-proxy/0.log" Dec 03 11:35:42 crc kubenswrapper[4646]: I1203 11:35:42.306480 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-z6m86_0be433f8-aae1-48a4-8853-de55ae48f836/manager/0.log" Dec 03 11:35:42 crc kubenswrapper[4646]: I1203 11:35:42.455186 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-hlwj7_e2fc2280-6814-4ba0-bd48-2272d37f2b41/manager/0.log" Dec 03 11:35:42 crc kubenswrapper[4646]: I1203 11:35:42.482155 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-hlwj7_e2fc2280-6814-4ba0-bd48-2272d37f2b41/kube-rbac-proxy/0.log" Dec 03 11:35:48 crc kubenswrapper[4646]: I1203 11:35:48.636340 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2lgzx"] Dec 03 11:35:48 crc kubenswrapper[4646]: E1203 11:35:48.637108 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="086eff50-3a10-4de4-8fd7-3ecfd54e5a25" containerName="container-00" Dec 03 11:35:48 crc kubenswrapper[4646]: I1203 11:35:48.637120 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="086eff50-3a10-4de4-8fd7-3ecfd54e5a25" containerName="container-00" Dec 03 11:35:48 crc kubenswrapper[4646]: I1203 11:35:48.683967 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="086eff50-3a10-4de4-8fd7-3ecfd54e5a25" containerName="container-00" Dec 03 11:35:48 crc kubenswrapper[4646]: I1203 11:35:48.687547 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:35:48 crc kubenswrapper[4646]: I1203 11:35:48.707545 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2lgzx"] Dec 03 11:35:48 crc kubenswrapper[4646]: I1203 11:35:48.820479 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74695981-c95d-4e95-b3f3-cba7da887dfd-catalog-content\") pod \"redhat-operators-2lgzx\" (UID: \"74695981-c95d-4e95-b3f3-cba7da887dfd\") " pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:35:48 crc kubenswrapper[4646]: I1203 11:35:48.820593 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwlxl\" (UniqueName: \"kubernetes.io/projected/74695981-c95d-4e95-b3f3-cba7da887dfd-kube-api-access-kwlxl\") pod \"redhat-operators-2lgzx\" (UID: \"74695981-c95d-4e95-b3f3-cba7da887dfd\") " pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:35:48 crc kubenswrapper[4646]: I1203 11:35:48.820673 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74695981-c95d-4e95-b3f3-cba7da887dfd-utilities\") pod \"redhat-operators-2lgzx\" (UID: \"74695981-c95d-4e95-b3f3-cba7da887dfd\") " pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:35:48 crc kubenswrapper[4646]: I1203 11:35:48.921910 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74695981-c95d-4e95-b3f3-cba7da887dfd-utilities\") pod \"redhat-operators-2lgzx\" (UID: \"74695981-c95d-4e95-b3f3-cba7da887dfd\") " pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:35:48 crc kubenswrapper[4646]: I1203 11:35:48.922535 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74695981-c95d-4e95-b3f3-cba7da887dfd-utilities\") pod \"redhat-operators-2lgzx\" (UID: \"74695981-c95d-4e95-b3f3-cba7da887dfd\") " pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:35:48 crc kubenswrapper[4646]: I1203 11:35:48.922554 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74695981-c95d-4e95-b3f3-cba7da887dfd-catalog-content\") pod \"redhat-operators-2lgzx\" (UID: \"74695981-c95d-4e95-b3f3-cba7da887dfd\") " pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:35:48 crc kubenswrapper[4646]: I1203 11:35:48.922870 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwlxl\" (UniqueName: \"kubernetes.io/projected/74695981-c95d-4e95-b3f3-cba7da887dfd-kube-api-access-kwlxl\") pod \"redhat-operators-2lgzx\" (UID: \"74695981-c95d-4e95-b3f3-cba7da887dfd\") " pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:35:48 crc kubenswrapper[4646]: I1203 11:35:48.923040 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74695981-c95d-4e95-b3f3-cba7da887dfd-catalog-content\") pod \"redhat-operators-2lgzx\" (UID: \"74695981-c95d-4e95-b3f3-cba7da887dfd\") " pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:35:48 crc kubenswrapper[4646]: I1203 11:35:48.958674 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwlxl\" (UniqueName: \"kubernetes.io/projected/74695981-c95d-4e95-b3f3-cba7da887dfd-kube-api-access-kwlxl\") pod \"redhat-operators-2lgzx\" (UID: \"74695981-c95d-4e95-b3f3-cba7da887dfd\") " pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:35:49 crc kubenswrapper[4646]: I1203 11:35:49.021306 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:35:49 crc kubenswrapper[4646]: I1203 11:35:49.566500 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2lgzx"] Dec 03 11:35:50 crc kubenswrapper[4646]: I1203 11:35:50.103876 4646 generic.go:334] "Generic (PLEG): container finished" podID="74695981-c95d-4e95-b3f3-cba7da887dfd" containerID="264630f153d5e9e2f6c05210946a959278f090e077712c1040f83f148e9ed517" exitCode=0 Dec 03 11:35:50 crc kubenswrapper[4646]: I1203 11:35:50.103925 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lgzx" event={"ID":"74695981-c95d-4e95-b3f3-cba7da887dfd","Type":"ContainerDied","Data":"264630f153d5e9e2f6c05210946a959278f090e077712c1040f83f148e9ed517"} Dec 03 11:35:50 crc kubenswrapper[4646]: I1203 11:35:50.104180 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lgzx" event={"ID":"74695981-c95d-4e95-b3f3-cba7da887dfd","Type":"ContainerStarted","Data":"b5cdeb1fb5aec327ef0029443b289a61e1a05ed5a3a08e32a635a2b447ab0dcc"} Dec 03 11:35:50 crc kubenswrapper[4646]: I1203 11:35:50.850110 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:35:50 crc kubenswrapper[4646]: E1203 11:35:50.850416 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:35:51 crc kubenswrapper[4646]: I1203 11:35:51.115826 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lgzx" event={"ID":"74695981-c95d-4e95-b3f3-cba7da887dfd","Type":"ContainerStarted","Data":"01923347e4177fcff5d3e732cb69310078e2338c7d1724b524b4cccb1361adfc"} Dec 03 11:35:55 crc kubenswrapper[4646]: I1203 11:35:55.146579 4646 generic.go:334] "Generic (PLEG): container finished" podID="74695981-c95d-4e95-b3f3-cba7da887dfd" containerID="01923347e4177fcff5d3e732cb69310078e2338c7d1724b524b4cccb1361adfc" exitCode=0 Dec 03 11:35:55 crc kubenswrapper[4646]: I1203 11:35:55.147121 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lgzx" event={"ID":"74695981-c95d-4e95-b3f3-cba7da887dfd","Type":"ContainerDied","Data":"01923347e4177fcff5d3e732cb69310078e2338c7d1724b524b4cccb1361adfc"} Dec 03 11:35:57 crc kubenswrapper[4646]: I1203 11:35:57.164148 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lgzx" event={"ID":"74695981-c95d-4e95-b3f3-cba7da887dfd","Type":"ContainerStarted","Data":"018e4c27f0456a4649a3d9503364093cbe051516c337ef8a58a174fffacb85f4"} Dec 03 11:35:57 crc kubenswrapper[4646]: I1203 11:35:57.191166 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2lgzx" podStartSLOduration=2.849505441 podStartE2EDuration="9.19114709s" podCreationTimestamp="2025-12-03 11:35:48 +0000 UTC" firstStartedPulling="2025-12-03 11:35:50.106838708 +0000 UTC m=+2526.569894843" lastFinishedPulling="2025-12-03 11:35:56.448480357 +0000 UTC m=+2532.911536492" observedRunningTime="2025-12-03 11:35:57.182303997 +0000 UTC m=+2533.645360132" watchObservedRunningTime="2025-12-03 11:35:57.19114709 +0000 UTC m=+2533.654203225" Dec 03 11:35:59 crc kubenswrapper[4646]: I1203 11:35:59.021617 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:35:59 crc kubenswrapper[4646]: I1203 11:35:59.021879 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:36:00 crc kubenswrapper[4646]: I1203 11:36:00.128756 4646 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2lgzx" podUID="74695981-c95d-4e95-b3f3-cba7da887dfd" containerName="registry-server" probeResult="failure" output=< Dec 03 11:36:00 crc kubenswrapper[4646]: timeout: failed to connect service ":50051" within 1s Dec 03 11:36:00 crc kubenswrapper[4646]: > Dec 03 11:36:01 crc kubenswrapper[4646]: I1203 11:36:01.849109 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:36:01 crc kubenswrapper[4646]: E1203 11:36:01.850033 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:36:05 crc kubenswrapper[4646]: I1203 11:36:05.743144 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-55nsv_041d1b1a-4861-4b7e-98ee-f5707404c173/control-plane-machine-set-operator/0.log" Dec 03 11:36:05 crc kubenswrapper[4646]: I1203 11:36:05.967831 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-lcfx6_1cc84826-c569-408e-94e9-8cf9bc49eae9/kube-rbac-proxy/0.log" Dec 03 11:36:05 crc kubenswrapper[4646]: I1203 11:36:05.971612 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-lcfx6_1cc84826-c569-408e-94e9-8cf9bc49eae9/machine-api-operator/0.log" Dec 03 11:36:09 crc kubenswrapper[4646]: I1203 11:36:09.074603 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:36:09 crc kubenswrapper[4646]: I1203 11:36:09.133565 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:36:09 crc kubenswrapper[4646]: I1203 11:36:09.307675 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2lgzx"] Dec 03 11:36:10 crc kubenswrapper[4646]: I1203 11:36:10.289736 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2lgzx" podUID="74695981-c95d-4e95-b3f3-cba7da887dfd" containerName="registry-server" containerID="cri-o://018e4c27f0456a4649a3d9503364093cbe051516c337ef8a58a174fffacb85f4" gracePeriod=2 Dec 03 11:36:10 crc kubenswrapper[4646]: I1203 11:36:10.738961 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:36:10 crc kubenswrapper[4646]: I1203 11:36:10.747568 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwlxl\" (UniqueName: \"kubernetes.io/projected/74695981-c95d-4e95-b3f3-cba7da887dfd-kube-api-access-kwlxl\") pod \"74695981-c95d-4e95-b3f3-cba7da887dfd\" (UID: \"74695981-c95d-4e95-b3f3-cba7da887dfd\") " Dec 03 11:36:10 crc kubenswrapper[4646]: I1203 11:36:10.747691 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74695981-c95d-4e95-b3f3-cba7da887dfd-utilities\") pod \"74695981-c95d-4e95-b3f3-cba7da887dfd\" (UID: \"74695981-c95d-4e95-b3f3-cba7da887dfd\") " Dec 03 11:36:10 crc kubenswrapper[4646]: I1203 11:36:10.747758 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74695981-c95d-4e95-b3f3-cba7da887dfd-catalog-content\") pod \"74695981-c95d-4e95-b3f3-cba7da887dfd\" (UID: \"74695981-c95d-4e95-b3f3-cba7da887dfd\") " Dec 03 11:36:10 crc kubenswrapper[4646]: I1203 11:36:10.748594 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74695981-c95d-4e95-b3f3-cba7da887dfd-utilities" (OuterVolumeSpecName: "utilities") pod "74695981-c95d-4e95-b3f3-cba7da887dfd" (UID: "74695981-c95d-4e95-b3f3-cba7da887dfd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:36:10 crc kubenswrapper[4646]: I1203 11:36:10.773798 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74695981-c95d-4e95-b3f3-cba7da887dfd-kube-api-access-kwlxl" (OuterVolumeSpecName: "kube-api-access-kwlxl") pod "74695981-c95d-4e95-b3f3-cba7da887dfd" (UID: "74695981-c95d-4e95-b3f3-cba7da887dfd"). InnerVolumeSpecName "kube-api-access-kwlxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:36:10 crc kubenswrapper[4646]: I1203 11:36:10.852379 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwlxl\" (UniqueName: \"kubernetes.io/projected/74695981-c95d-4e95-b3f3-cba7da887dfd-kube-api-access-kwlxl\") on node \"crc\" DevicePath \"\"" Dec 03 11:36:10 crc kubenswrapper[4646]: I1203 11:36:10.852698 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74695981-c95d-4e95-b3f3-cba7da887dfd-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 11:36:10 crc kubenswrapper[4646]: I1203 11:36:10.912455 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74695981-c95d-4e95-b3f3-cba7da887dfd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74695981-c95d-4e95-b3f3-cba7da887dfd" (UID: "74695981-c95d-4e95-b3f3-cba7da887dfd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:36:10 crc kubenswrapper[4646]: I1203 11:36:10.954768 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74695981-c95d-4e95-b3f3-cba7da887dfd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.301394 4646 generic.go:334] "Generic (PLEG): container finished" podID="74695981-c95d-4e95-b3f3-cba7da887dfd" containerID="018e4c27f0456a4649a3d9503364093cbe051516c337ef8a58a174fffacb85f4" exitCode=0 Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.301442 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lgzx" event={"ID":"74695981-c95d-4e95-b3f3-cba7da887dfd","Type":"ContainerDied","Data":"018e4c27f0456a4649a3d9503364093cbe051516c337ef8a58a174fffacb85f4"} Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.301475 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lgzx" event={"ID":"74695981-c95d-4e95-b3f3-cba7da887dfd","Type":"ContainerDied","Data":"b5cdeb1fb5aec327ef0029443b289a61e1a05ed5a3a08e32a635a2b447ab0dcc"} Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.301497 4646 scope.go:117] "RemoveContainer" containerID="018e4c27f0456a4649a3d9503364093cbe051516c337ef8a58a174fffacb85f4" Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.301654 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2lgzx" Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.317610 4646 scope.go:117] "RemoveContainer" containerID="01923347e4177fcff5d3e732cb69310078e2338c7d1724b524b4cccb1361adfc" Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.334715 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2lgzx"] Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.358541 4646 scope.go:117] "RemoveContainer" containerID="264630f153d5e9e2f6c05210946a959278f090e077712c1040f83f148e9ed517" Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.359658 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2lgzx"] Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.390212 4646 scope.go:117] "RemoveContainer" containerID="018e4c27f0456a4649a3d9503364093cbe051516c337ef8a58a174fffacb85f4" Dec 03 11:36:11 crc kubenswrapper[4646]: E1203 11:36:11.390838 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"018e4c27f0456a4649a3d9503364093cbe051516c337ef8a58a174fffacb85f4\": container with ID starting with 018e4c27f0456a4649a3d9503364093cbe051516c337ef8a58a174fffacb85f4 not found: ID does not exist" containerID="018e4c27f0456a4649a3d9503364093cbe051516c337ef8a58a174fffacb85f4" Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.390903 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"018e4c27f0456a4649a3d9503364093cbe051516c337ef8a58a174fffacb85f4"} err="failed to get container status \"018e4c27f0456a4649a3d9503364093cbe051516c337ef8a58a174fffacb85f4\": rpc error: code = NotFound desc = could not find container \"018e4c27f0456a4649a3d9503364093cbe051516c337ef8a58a174fffacb85f4\": container with ID starting with 018e4c27f0456a4649a3d9503364093cbe051516c337ef8a58a174fffacb85f4 not found: ID does not exist" Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.390930 4646 scope.go:117] "RemoveContainer" containerID="01923347e4177fcff5d3e732cb69310078e2338c7d1724b524b4cccb1361adfc" Dec 03 11:36:11 crc kubenswrapper[4646]: E1203 11:36:11.391557 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01923347e4177fcff5d3e732cb69310078e2338c7d1724b524b4cccb1361adfc\": container with ID starting with 01923347e4177fcff5d3e732cb69310078e2338c7d1724b524b4cccb1361adfc not found: ID does not exist" containerID="01923347e4177fcff5d3e732cb69310078e2338c7d1724b524b4cccb1361adfc" Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.391619 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01923347e4177fcff5d3e732cb69310078e2338c7d1724b524b4cccb1361adfc"} err="failed to get container status \"01923347e4177fcff5d3e732cb69310078e2338c7d1724b524b4cccb1361adfc\": rpc error: code = NotFound desc = could not find container \"01923347e4177fcff5d3e732cb69310078e2338c7d1724b524b4cccb1361adfc\": container with ID starting with 01923347e4177fcff5d3e732cb69310078e2338c7d1724b524b4cccb1361adfc not found: ID does not exist" Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.391639 4646 scope.go:117] "RemoveContainer" containerID="264630f153d5e9e2f6c05210946a959278f090e077712c1040f83f148e9ed517" Dec 03 11:36:11 crc kubenswrapper[4646]: E1203 11:36:11.391856 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"264630f153d5e9e2f6c05210946a959278f090e077712c1040f83f148e9ed517\": container with ID starting with 264630f153d5e9e2f6c05210946a959278f090e077712c1040f83f148e9ed517 not found: ID does not exist" containerID="264630f153d5e9e2f6c05210946a959278f090e077712c1040f83f148e9ed517" Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.391882 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"264630f153d5e9e2f6c05210946a959278f090e077712c1040f83f148e9ed517"} err="failed to get container status \"264630f153d5e9e2f6c05210946a959278f090e077712c1040f83f148e9ed517\": rpc error: code = NotFound desc = could not find container \"264630f153d5e9e2f6c05210946a959278f090e077712c1040f83f148e9ed517\": container with ID starting with 264630f153d5e9e2f6c05210946a959278f090e077712c1040f83f148e9ed517 not found: ID does not exist" Dec 03 11:36:11 crc kubenswrapper[4646]: I1203 11:36:11.861122 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74695981-c95d-4e95-b3f3-cba7da887dfd" path="/var/lib/kubelet/pods/74695981-c95d-4e95-b3f3-cba7da887dfd/volumes" Dec 03 11:36:15 crc kubenswrapper[4646]: I1203 11:36:15.849935 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:36:15 crc kubenswrapper[4646]: E1203 11:36:15.850706 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:36:19 crc kubenswrapper[4646]: I1203 11:36:19.075401 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-xpd69_22a8bf20-667c-4d61-8030-0cd69dbb0d37/cert-manager-controller/0.log" Dec 03 11:36:19 crc kubenswrapper[4646]: I1203 11:36:19.225967 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-96cjs_345d3005-d281-404b-92bf-7f2613a262f8/cert-manager-cainjector/0.log" Dec 03 11:36:19 crc kubenswrapper[4646]: I1203 11:36:19.283286 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-2n8gn_8a2565ee-c652-4918-8ca5-c114ce33fe17/cert-manager-webhook/0.log" Dec 03 11:36:28 crc kubenswrapper[4646]: I1203 11:36:28.848005 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:36:28 crc kubenswrapper[4646]: E1203 11:36:28.848955 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:36:31 crc kubenswrapper[4646]: I1203 11:36:31.276324 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-nds5d_277597ef-bb30-4377-ba72-7da7a58d8d52/nmstate-console-plugin/0.log" Dec 03 11:36:31 crc kubenswrapper[4646]: I1203 11:36:31.479663 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-mpgkf_fe49c8ca-0ee9-436e-bf21-dda2e34d2032/kube-rbac-proxy/0.log" Dec 03 11:36:31 crc kubenswrapper[4646]: I1203 11:36:31.480035 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-j8lb8_2c460b1c-cc8b-4455-9ae5-ee778c542705/nmstate-handler/0.log" Dec 03 11:36:31 crc kubenswrapper[4646]: I1203 11:36:31.525464 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-mpgkf_fe49c8ca-0ee9-436e-bf21-dda2e34d2032/nmstate-metrics/0.log" Dec 03 11:36:31 crc kubenswrapper[4646]: I1203 11:36:31.702074 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-5pn6t_9094b6b6-398b-4d47-96c5-45c236227b1c/nmstate-operator/0.log" Dec 03 11:36:31 crc kubenswrapper[4646]: I1203 11:36:31.793040 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-7dflq_1a81e55c-4071-4cd3-8e5d-5d51477c994b/nmstate-webhook/0.log" Dec 03 11:36:39 crc kubenswrapper[4646]: I1203 11:36:39.848794 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:36:39 crc kubenswrapper[4646]: E1203 11:36:39.849610 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:36:45 crc kubenswrapper[4646]: I1203 11:36:45.989403 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-jfnxz_d6afebba-4203-45bd-831c-f9d399d27957/kube-rbac-proxy/0.log" Dec 03 11:36:46 crc kubenswrapper[4646]: I1203 11:36:46.164966 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-jfnxz_d6afebba-4203-45bd-831c-f9d399d27957/controller/0.log" Dec 03 11:36:46 crc kubenswrapper[4646]: I1203 11:36:46.239438 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/cp-frr-files/0.log" Dec 03 11:36:46 crc kubenswrapper[4646]: I1203 11:36:46.434579 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/cp-reloader/0.log" Dec 03 11:36:46 crc kubenswrapper[4646]: I1203 11:36:46.452738 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/cp-frr-files/0.log" Dec 03 11:36:46 crc kubenswrapper[4646]: I1203 11:36:46.473890 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/cp-reloader/0.log" Dec 03 11:36:46 crc kubenswrapper[4646]: I1203 11:36:46.492807 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/cp-metrics/0.log" Dec 03 11:36:46 crc kubenswrapper[4646]: I1203 11:36:46.721436 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/cp-frr-files/0.log" Dec 03 11:36:46 crc kubenswrapper[4646]: I1203 11:36:46.732448 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/cp-metrics/0.log" Dec 03 11:36:46 crc kubenswrapper[4646]: I1203 11:36:46.733118 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/cp-reloader/0.log" Dec 03 11:36:46 crc kubenswrapper[4646]: I1203 11:36:46.772083 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/cp-metrics/0.log" Dec 03 11:36:46 crc kubenswrapper[4646]: I1203 11:36:46.927725 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/cp-frr-files/0.log" Dec 03 11:36:46 crc kubenswrapper[4646]: I1203 11:36:46.980133 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/cp-reloader/0.log" Dec 03 11:36:46 crc kubenswrapper[4646]: I1203 11:36:46.997964 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/cp-metrics/0.log" Dec 03 11:36:47 crc kubenswrapper[4646]: I1203 11:36:47.003619 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/controller/0.log" Dec 03 11:36:47 crc kubenswrapper[4646]: I1203 11:36:47.201721 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/kube-rbac-proxy-frr/0.log" Dec 03 11:36:47 crc kubenswrapper[4646]: I1203 11:36:47.223865 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/kube-rbac-proxy/0.log" Dec 03 11:36:47 crc kubenswrapper[4646]: I1203 11:36:47.225471 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/frr-metrics/0.log" Dec 03 11:36:47 crc kubenswrapper[4646]: I1203 11:36:47.445559 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/reloader/0.log" Dec 03 11:36:47 crc kubenswrapper[4646]: I1203 11:36:47.626005 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-2ff9f_c5af6439-97a5-487d-b4a2-0fd0660af159/frr-k8s-webhook-server/0.log" Dec 03 11:36:47 crc kubenswrapper[4646]: I1203 11:36:47.855314 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-74454568d4-nvshz_1d60bcfb-69b3-4552-9610-e4f936439116/manager/0.log" Dec 03 11:36:48 crc kubenswrapper[4646]: I1203 11:36:48.068005 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-knlr4_a44776c2-244d-4ba3-b2f5-323eef5c8b9f/frr/0.log" Dec 03 11:36:48 crc kubenswrapper[4646]: I1203 11:36:48.092302 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6c975599d9-tq9v4_70261b97-2e67-42c1-be85-44df43cf845d/webhook-server/0.log" Dec 03 11:36:48 crc kubenswrapper[4646]: I1203 11:36:48.202727 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-vxhmd_72df546c-4ffc-46d9-9f79-54bb788a08fc/kube-rbac-proxy/0.log" Dec 03 11:36:48 crc kubenswrapper[4646]: I1203 11:36:48.713866 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-vxhmd_72df546c-4ffc-46d9-9f79-54bb788a08fc/speaker/0.log" Dec 03 11:36:54 crc kubenswrapper[4646]: I1203 11:36:54.848675 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:36:54 crc kubenswrapper[4646]: E1203 11:36:54.849500 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.696269 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6lwcq"] Dec 03 11:36:55 crc kubenswrapper[4646]: E1203 11:36:55.697442 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74695981-c95d-4e95-b3f3-cba7da887dfd" containerName="registry-server" Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.697551 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="74695981-c95d-4e95-b3f3-cba7da887dfd" containerName="registry-server" Dec 03 11:36:55 crc kubenswrapper[4646]: E1203 11:36:55.697643 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74695981-c95d-4e95-b3f3-cba7da887dfd" containerName="extract-utilities" Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.698081 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="74695981-c95d-4e95-b3f3-cba7da887dfd" containerName="extract-utilities" Dec 03 11:36:55 crc kubenswrapper[4646]: E1203 11:36:55.698182 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74695981-c95d-4e95-b3f3-cba7da887dfd" containerName="extract-content" Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.698237 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="74695981-c95d-4e95-b3f3-cba7da887dfd" containerName="extract-content" Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.698520 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="74695981-c95d-4e95-b3f3-cba7da887dfd" containerName="registry-server" Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.701679 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.719885 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6lwcq"] Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.781549 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38dffc40-f43f-46d4-88ac-8084230ece0f-utilities\") pod \"certified-operators-6lwcq\" (UID: \"38dffc40-f43f-46d4-88ac-8084230ece0f\") " pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.781963 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38dffc40-f43f-46d4-88ac-8084230ece0f-catalog-content\") pod \"certified-operators-6lwcq\" (UID: \"38dffc40-f43f-46d4-88ac-8084230ece0f\") " pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.782024 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2tv5\" (UniqueName: \"kubernetes.io/projected/38dffc40-f43f-46d4-88ac-8084230ece0f-kube-api-access-t2tv5\") pod \"certified-operators-6lwcq\" (UID: \"38dffc40-f43f-46d4-88ac-8084230ece0f\") " pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.883427 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38dffc40-f43f-46d4-88ac-8084230ece0f-utilities\") pod \"certified-operators-6lwcq\" (UID: \"38dffc40-f43f-46d4-88ac-8084230ece0f\") " pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.883498 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38dffc40-f43f-46d4-88ac-8084230ece0f-catalog-content\") pod \"certified-operators-6lwcq\" (UID: \"38dffc40-f43f-46d4-88ac-8084230ece0f\") " pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.883531 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2tv5\" (UniqueName: \"kubernetes.io/projected/38dffc40-f43f-46d4-88ac-8084230ece0f-kube-api-access-t2tv5\") pod \"certified-operators-6lwcq\" (UID: \"38dffc40-f43f-46d4-88ac-8084230ece0f\") " pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.884045 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38dffc40-f43f-46d4-88ac-8084230ece0f-catalog-content\") pod \"certified-operators-6lwcq\" (UID: \"38dffc40-f43f-46d4-88ac-8084230ece0f\") " pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.884045 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38dffc40-f43f-46d4-88ac-8084230ece0f-utilities\") pod \"certified-operators-6lwcq\" (UID: \"38dffc40-f43f-46d4-88ac-8084230ece0f\") " pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:36:55 crc kubenswrapper[4646]: I1203 11:36:55.919315 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2tv5\" (UniqueName: \"kubernetes.io/projected/38dffc40-f43f-46d4-88ac-8084230ece0f-kube-api-access-t2tv5\") pod \"certified-operators-6lwcq\" (UID: \"38dffc40-f43f-46d4-88ac-8084230ece0f\") " pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:36:56 crc kubenswrapper[4646]: I1203 11:36:56.029115 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:36:56 crc kubenswrapper[4646]: I1203 11:36:56.481770 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6lwcq"] Dec 03 11:36:56 crc kubenswrapper[4646]: I1203 11:36:56.729449 4646 generic.go:334] "Generic (PLEG): container finished" podID="38dffc40-f43f-46d4-88ac-8084230ece0f" containerID="a2c7a68aea18783bab9fde0216f7401f5b3266742be113e22f293df0a5b3a33a" exitCode=0 Dec 03 11:36:56 crc kubenswrapper[4646]: I1203 11:36:56.729608 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lwcq" event={"ID":"38dffc40-f43f-46d4-88ac-8084230ece0f","Type":"ContainerDied","Data":"a2c7a68aea18783bab9fde0216f7401f5b3266742be113e22f293df0a5b3a33a"} Dec 03 11:36:56 crc kubenswrapper[4646]: I1203 11:36:56.729722 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lwcq" event={"ID":"38dffc40-f43f-46d4-88ac-8084230ece0f","Type":"ContainerStarted","Data":"8547a91a47d4a5ac50afb51c5ac352772a68825df69a50a748cbe5190b8a4bde"} Dec 03 11:36:56 crc kubenswrapper[4646]: I1203 11:36:56.732673 4646 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 11:36:57 crc kubenswrapper[4646]: I1203 11:36:57.738230 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lwcq" event={"ID":"38dffc40-f43f-46d4-88ac-8084230ece0f","Type":"ContainerStarted","Data":"441d6e0bd2186e684799d3840de34053cb99658b4ed4ec34863832378710987e"} Dec 03 11:36:59 crc kubenswrapper[4646]: I1203 11:36:59.757834 4646 generic.go:334] "Generic (PLEG): container finished" podID="38dffc40-f43f-46d4-88ac-8084230ece0f" containerID="441d6e0bd2186e684799d3840de34053cb99658b4ed4ec34863832378710987e" exitCode=0 Dec 03 11:36:59 crc kubenswrapper[4646]: I1203 11:36:59.757938 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lwcq" event={"ID":"38dffc40-f43f-46d4-88ac-8084230ece0f","Type":"ContainerDied","Data":"441d6e0bd2186e684799d3840de34053cb99658b4ed4ec34863832378710987e"} Dec 03 11:37:00 crc kubenswrapper[4646]: I1203 11:37:00.770729 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lwcq" event={"ID":"38dffc40-f43f-46d4-88ac-8084230ece0f","Type":"ContainerStarted","Data":"9a43eb63728a5286f653ffbb4a788d4562ad0e69f8a2af4a9da90b39d7f28158"} Dec 03 11:37:00 crc kubenswrapper[4646]: I1203 11:37:00.796385 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6lwcq" podStartSLOduration=2.22768343 podStartE2EDuration="5.796367759s" podCreationTimestamp="2025-12-03 11:36:55 +0000 UTC" firstStartedPulling="2025-12-03 11:36:56.732373883 +0000 UTC m=+2593.195430018" lastFinishedPulling="2025-12-03 11:37:00.301058212 +0000 UTC m=+2596.764114347" observedRunningTime="2025-12-03 11:37:00.795538225 +0000 UTC m=+2597.258594370" watchObservedRunningTime="2025-12-03 11:37:00.796367759 +0000 UTC m=+2597.259423894" Dec 03 11:37:02 crc kubenswrapper[4646]: I1203 11:37:02.643146 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc_d37842a3-b44d-4fa4-a67a-7e1888ed8832/util/0.log" Dec 03 11:37:02 crc kubenswrapper[4646]: I1203 11:37:02.862348 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc_d37842a3-b44d-4fa4-a67a-7e1888ed8832/pull/0.log" Dec 03 11:37:02 crc kubenswrapper[4646]: I1203 11:37:02.922712 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc_d37842a3-b44d-4fa4-a67a-7e1888ed8832/pull/0.log" Dec 03 11:37:02 crc kubenswrapper[4646]: I1203 11:37:02.936154 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc_d37842a3-b44d-4fa4-a67a-7e1888ed8832/util/0.log" Dec 03 11:37:03 crc kubenswrapper[4646]: I1203 11:37:03.067178 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc_d37842a3-b44d-4fa4-a67a-7e1888ed8832/util/0.log" Dec 03 11:37:03 crc kubenswrapper[4646]: I1203 11:37:03.152961 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc_d37842a3-b44d-4fa4-a67a-7e1888ed8832/extract/0.log" Dec 03 11:37:03 crc kubenswrapper[4646]: I1203 11:37:03.232521 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn8pcc_d37842a3-b44d-4fa4-a67a-7e1888ed8832/pull/0.log" Dec 03 11:37:03 crc kubenswrapper[4646]: I1203 11:37:03.302000 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q_d8fd2b49-2283-4290-8e40-9d14ad64ca72/util/0.log" Dec 03 11:37:03 crc kubenswrapper[4646]: I1203 11:37:03.566604 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q_d8fd2b49-2283-4290-8e40-9d14ad64ca72/util/0.log" Dec 03 11:37:03 crc kubenswrapper[4646]: I1203 11:37:03.663971 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q_d8fd2b49-2283-4290-8e40-9d14ad64ca72/pull/0.log" Dec 03 11:37:03 crc kubenswrapper[4646]: I1203 11:37:03.664020 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q_d8fd2b49-2283-4290-8e40-9d14ad64ca72/pull/0.log" Dec 03 11:37:03 crc kubenswrapper[4646]: I1203 11:37:03.749733 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q_d8fd2b49-2283-4290-8e40-9d14ad64ca72/util/0.log" Dec 03 11:37:03 crc kubenswrapper[4646]: I1203 11:37:03.780151 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q_d8fd2b49-2283-4290-8e40-9d14ad64ca72/pull/0.log" Dec 03 11:37:03 crc kubenswrapper[4646]: I1203 11:37:03.953934 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lwcq_38dffc40-f43f-46d4-88ac-8084230ece0f/extract-utilities/0.log" Dec 03 11:37:04 crc kubenswrapper[4646]: I1203 11:37:04.205721 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lwcq_38dffc40-f43f-46d4-88ac-8084230ece0f/extract-content/0.log" Dec 03 11:37:04 crc kubenswrapper[4646]: I1203 11:37:04.253326 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lwcq_38dffc40-f43f-46d4-88ac-8084230ece0f/extract-utilities/0.log" Dec 03 11:37:04 crc kubenswrapper[4646]: I1203 11:37:04.262096 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lwcq_38dffc40-f43f-46d4-88ac-8084230ece0f/extract-content/0.log" Dec 03 11:37:04 crc kubenswrapper[4646]: I1203 11:37:04.321053 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fjb2q_d8fd2b49-2283-4290-8e40-9d14ad64ca72/extract/0.log" Dec 03 11:37:04 crc kubenswrapper[4646]: I1203 11:37:04.436421 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lwcq_38dffc40-f43f-46d4-88ac-8084230ece0f/extract-utilities/0.log" Dec 03 11:37:04 crc kubenswrapper[4646]: I1203 11:37:04.475020 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lwcq_38dffc40-f43f-46d4-88ac-8084230ece0f/extract-content/0.log" Dec 03 11:37:04 crc kubenswrapper[4646]: I1203 11:37:04.504231 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lwcq_38dffc40-f43f-46d4-88ac-8084230ece0f/registry-server/0.log" Dec 03 11:37:04 crc kubenswrapper[4646]: I1203 11:37:04.679934 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qspkq_682649bd-4003-4e88-a81a-dafee3703114/extract-utilities/0.log" Dec 03 11:37:04 crc kubenswrapper[4646]: I1203 11:37:04.958572 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qspkq_682649bd-4003-4e88-a81a-dafee3703114/extract-content/0.log" Dec 03 11:37:04 crc kubenswrapper[4646]: I1203 11:37:04.962799 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qspkq_682649bd-4003-4e88-a81a-dafee3703114/extract-utilities/0.log" Dec 03 11:37:04 crc kubenswrapper[4646]: I1203 11:37:04.970387 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qspkq_682649bd-4003-4e88-a81a-dafee3703114/extract-content/0.log" Dec 03 11:37:05 crc kubenswrapper[4646]: I1203 11:37:05.261464 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qspkq_682649bd-4003-4e88-a81a-dafee3703114/extract-content/0.log" Dec 03 11:37:05 crc kubenswrapper[4646]: I1203 11:37:05.541887 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qspkq_682649bd-4003-4e88-a81a-dafee3703114/registry-server/0.log" Dec 03 11:37:05 crc kubenswrapper[4646]: I1203 11:37:05.542577 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kbtw5_9ccf8b15-522e-4425-9ce3-8ba8721454b7/extract-utilities/0.log" Dec 03 11:37:05 crc kubenswrapper[4646]: I1203 11:37:05.612687 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qspkq_682649bd-4003-4e88-a81a-dafee3703114/extract-utilities/0.log" Dec 03 11:37:06 crc kubenswrapper[4646]: I1203 11:37:06.032271 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:37:06 crc kubenswrapper[4646]: I1203 11:37:06.032384 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:37:06 crc kubenswrapper[4646]: I1203 11:37:06.049147 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kbtw5_9ccf8b15-522e-4425-9ce3-8ba8721454b7/extract-content/0.log" Dec 03 11:37:06 crc kubenswrapper[4646]: I1203 11:37:06.096242 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kbtw5_9ccf8b15-522e-4425-9ce3-8ba8721454b7/extract-content/0.log" Dec 03 11:37:06 crc kubenswrapper[4646]: I1203 11:37:06.099310 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:37:06 crc kubenswrapper[4646]: I1203 11:37:06.130100 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kbtw5_9ccf8b15-522e-4425-9ce3-8ba8721454b7/extract-utilities/0.log" Dec 03 11:37:06 crc kubenswrapper[4646]: I1203 11:37:06.421863 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kbtw5_9ccf8b15-522e-4425-9ce3-8ba8721454b7/extract-content/0.log" Dec 03 11:37:06 crc kubenswrapper[4646]: I1203 11:37:06.443867 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kbtw5_9ccf8b15-522e-4425-9ce3-8ba8721454b7/extract-utilities/0.log" Dec 03 11:37:06 crc kubenswrapper[4646]: I1203 11:37:06.660639 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-qvvn5_53441ce7-95f4-4236-8df4-fc04de470569/marketplace-operator/0.log" Dec 03 11:37:06 crc kubenswrapper[4646]: I1203 11:37:06.789259 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dtn6s_5394e3ff-06b8-4220-81c1-21475d7769fa/extract-utilities/0.log" Dec 03 11:37:06 crc kubenswrapper[4646]: I1203 11:37:06.897723 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:37:06 crc kubenswrapper[4646]: I1203 11:37:06.970743 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6lwcq"] Dec 03 11:37:06 crc kubenswrapper[4646]: I1203 11:37:06.989255 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kbtw5_9ccf8b15-522e-4425-9ce3-8ba8721454b7/registry-server/0.log" Dec 03 11:37:07 crc kubenswrapper[4646]: I1203 11:37:07.364542 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dtn6s_5394e3ff-06b8-4220-81c1-21475d7769fa/extract-content/0.log" Dec 03 11:37:07 crc kubenswrapper[4646]: I1203 11:37:07.408783 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dtn6s_5394e3ff-06b8-4220-81c1-21475d7769fa/extract-utilities/0.log" Dec 03 11:37:07 crc kubenswrapper[4646]: I1203 11:37:07.419858 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dtn6s_5394e3ff-06b8-4220-81c1-21475d7769fa/extract-content/0.log" Dec 03 11:37:07 crc kubenswrapper[4646]: I1203 11:37:07.594498 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dtn6s_5394e3ff-06b8-4220-81c1-21475d7769fa/extract-content/0.log" Dec 03 11:37:07 crc kubenswrapper[4646]: I1203 11:37:07.649975 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-82b96_0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e/extract-utilities/0.log" Dec 03 11:37:07 crc kubenswrapper[4646]: I1203 11:37:07.720620 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dtn6s_5394e3ff-06b8-4220-81c1-21475d7769fa/registry-server/0.log" Dec 03 11:37:07 crc kubenswrapper[4646]: I1203 11:37:07.739450 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dtn6s_5394e3ff-06b8-4220-81c1-21475d7769fa/extract-utilities/0.log" Dec 03 11:37:07 crc kubenswrapper[4646]: I1203 11:37:07.943374 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-82b96_0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e/extract-content/0.log" Dec 03 11:37:07 crc kubenswrapper[4646]: I1203 11:37:07.944995 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-82b96_0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e/extract-utilities/0.log" Dec 03 11:37:07 crc kubenswrapper[4646]: I1203 11:37:07.997731 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-82b96_0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e/extract-content/0.log" Dec 03 11:37:08 crc kubenswrapper[4646]: I1203 11:37:08.289743 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-82b96_0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e/extract-content/0.log" Dec 03 11:37:08 crc kubenswrapper[4646]: I1203 11:37:08.310757 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-82b96_0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e/extract-utilities/0.log" Dec 03 11:37:08 crc kubenswrapper[4646]: I1203 11:37:08.611565 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-82b96_0bc4f0c6-82a4-4b75-bf15-3f4e5a4f360e/registry-server/0.log" Dec 03 11:37:08 crc kubenswrapper[4646]: I1203 11:37:08.851724 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6lwcq" podUID="38dffc40-f43f-46d4-88ac-8084230ece0f" containerName="registry-server" containerID="cri-o://9a43eb63728a5286f653ffbb4a788d4562ad0e69f8a2af4a9da90b39d7f28158" gracePeriod=2 Dec 03 11:37:08 crc kubenswrapper[4646]: I1203 11:37:08.855258 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:37:08 crc kubenswrapper[4646]: E1203 11:37:08.855611 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.302105 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.408850 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38dffc40-f43f-46d4-88ac-8084230ece0f-catalog-content\") pod \"38dffc40-f43f-46d4-88ac-8084230ece0f\" (UID: \"38dffc40-f43f-46d4-88ac-8084230ece0f\") " Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.408956 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2tv5\" (UniqueName: \"kubernetes.io/projected/38dffc40-f43f-46d4-88ac-8084230ece0f-kube-api-access-t2tv5\") pod \"38dffc40-f43f-46d4-88ac-8084230ece0f\" (UID: \"38dffc40-f43f-46d4-88ac-8084230ece0f\") " Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.409044 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38dffc40-f43f-46d4-88ac-8084230ece0f-utilities\") pod \"38dffc40-f43f-46d4-88ac-8084230ece0f\" (UID: \"38dffc40-f43f-46d4-88ac-8084230ece0f\") " Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.410358 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38dffc40-f43f-46d4-88ac-8084230ece0f-utilities" (OuterVolumeSpecName: "utilities") pod "38dffc40-f43f-46d4-88ac-8084230ece0f" (UID: "38dffc40-f43f-46d4-88ac-8084230ece0f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.414455 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38dffc40-f43f-46d4-88ac-8084230ece0f-kube-api-access-t2tv5" (OuterVolumeSpecName: "kube-api-access-t2tv5") pod "38dffc40-f43f-46d4-88ac-8084230ece0f" (UID: "38dffc40-f43f-46d4-88ac-8084230ece0f"). InnerVolumeSpecName "kube-api-access-t2tv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.511987 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2tv5\" (UniqueName: \"kubernetes.io/projected/38dffc40-f43f-46d4-88ac-8084230ece0f-kube-api-access-t2tv5\") on node \"crc\" DevicePath \"\"" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.512028 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38dffc40-f43f-46d4-88ac-8084230ece0f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.535751 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38dffc40-f43f-46d4-88ac-8084230ece0f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38dffc40-f43f-46d4-88ac-8084230ece0f" (UID: "38dffc40-f43f-46d4-88ac-8084230ece0f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.613512 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38dffc40-f43f-46d4-88ac-8084230ece0f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.862477 4646 generic.go:334] "Generic (PLEG): container finished" podID="38dffc40-f43f-46d4-88ac-8084230ece0f" containerID="9a43eb63728a5286f653ffbb4a788d4562ad0e69f8a2af4a9da90b39d7f28158" exitCode=0 Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.862518 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lwcq" event={"ID":"38dffc40-f43f-46d4-88ac-8084230ece0f","Type":"ContainerDied","Data":"9a43eb63728a5286f653ffbb4a788d4562ad0e69f8a2af4a9da90b39d7f28158"} Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.862574 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6lwcq" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.863003 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lwcq" event={"ID":"38dffc40-f43f-46d4-88ac-8084230ece0f","Type":"ContainerDied","Data":"8547a91a47d4a5ac50afb51c5ac352772a68825df69a50a748cbe5190b8a4bde"} Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.863090 4646 scope.go:117] "RemoveContainer" containerID="9a43eb63728a5286f653ffbb4a788d4562ad0e69f8a2af4a9da90b39d7f28158" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.890156 4646 scope.go:117] "RemoveContainer" containerID="441d6e0bd2186e684799d3840de34053cb99658b4ed4ec34863832378710987e" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.918658 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6lwcq"] Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.942535 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6lwcq"] Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.943058 4646 scope.go:117] "RemoveContainer" containerID="a2c7a68aea18783bab9fde0216f7401f5b3266742be113e22f293df0a5b3a33a" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.959638 4646 scope.go:117] "RemoveContainer" containerID="9a43eb63728a5286f653ffbb4a788d4562ad0e69f8a2af4a9da90b39d7f28158" Dec 03 11:37:09 crc kubenswrapper[4646]: E1203 11:37:09.960058 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a43eb63728a5286f653ffbb4a788d4562ad0e69f8a2af4a9da90b39d7f28158\": container with ID starting with 9a43eb63728a5286f653ffbb4a788d4562ad0e69f8a2af4a9da90b39d7f28158 not found: ID does not exist" containerID="9a43eb63728a5286f653ffbb4a788d4562ad0e69f8a2af4a9da90b39d7f28158" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.960111 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a43eb63728a5286f653ffbb4a788d4562ad0e69f8a2af4a9da90b39d7f28158"} err="failed to get container status \"9a43eb63728a5286f653ffbb4a788d4562ad0e69f8a2af4a9da90b39d7f28158\": rpc error: code = NotFound desc = could not find container \"9a43eb63728a5286f653ffbb4a788d4562ad0e69f8a2af4a9da90b39d7f28158\": container with ID starting with 9a43eb63728a5286f653ffbb4a788d4562ad0e69f8a2af4a9da90b39d7f28158 not found: ID does not exist" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.960132 4646 scope.go:117] "RemoveContainer" containerID="441d6e0bd2186e684799d3840de34053cb99658b4ed4ec34863832378710987e" Dec 03 11:37:09 crc kubenswrapper[4646]: E1203 11:37:09.972246 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"441d6e0bd2186e684799d3840de34053cb99658b4ed4ec34863832378710987e\": container with ID starting with 441d6e0bd2186e684799d3840de34053cb99658b4ed4ec34863832378710987e not found: ID does not exist" containerID="441d6e0bd2186e684799d3840de34053cb99658b4ed4ec34863832378710987e" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.972292 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"441d6e0bd2186e684799d3840de34053cb99658b4ed4ec34863832378710987e"} err="failed to get container status \"441d6e0bd2186e684799d3840de34053cb99658b4ed4ec34863832378710987e\": rpc error: code = NotFound desc = could not find container \"441d6e0bd2186e684799d3840de34053cb99658b4ed4ec34863832378710987e\": container with ID starting with 441d6e0bd2186e684799d3840de34053cb99658b4ed4ec34863832378710987e not found: ID does not exist" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.972326 4646 scope.go:117] "RemoveContainer" containerID="a2c7a68aea18783bab9fde0216f7401f5b3266742be113e22f293df0a5b3a33a" Dec 03 11:37:09 crc kubenswrapper[4646]: E1203 11:37:09.972877 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2c7a68aea18783bab9fde0216f7401f5b3266742be113e22f293df0a5b3a33a\": container with ID starting with a2c7a68aea18783bab9fde0216f7401f5b3266742be113e22f293df0a5b3a33a not found: ID does not exist" containerID="a2c7a68aea18783bab9fde0216f7401f5b3266742be113e22f293df0a5b3a33a" Dec 03 11:37:09 crc kubenswrapper[4646]: I1203 11:37:09.972919 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2c7a68aea18783bab9fde0216f7401f5b3266742be113e22f293df0a5b3a33a"} err="failed to get container status \"a2c7a68aea18783bab9fde0216f7401f5b3266742be113e22f293df0a5b3a33a\": rpc error: code = NotFound desc = could not find container \"a2c7a68aea18783bab9fde0216f7401f5b3266742be113e22f293df0a5b3a33a\": container with ID starting with a2c7a68aea18783bab9fde0216f7401f5b3266742be113e22f293df0a5b3a33a not found: ID does not exist" Dec 03 11:37:11 crc kubenswrapper[4646]: I1203 11:37:11.859368 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38dffc40-f43f-46d4-88ac-8084230ece0f" path="/var/lib/kubelet/pods/38dffc40-f43f-46d4-88ac-8084230ece0f/volumes" Dec 03 11:37:22 crc kubenswrapper[4646]: I1203 11:37:22.848326 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:37:22 crc kubenswrapper[4646]: E1203 11:37:22.849083 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:37:37 crc kubenswrapper[4646]: I1203 11:37:37.848430 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:37:37 crc kubenswrapper[4646]: E1203 11:37:37.849272 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:37:52 crc kubenswrapper[4646]: I1203 11:37:52.848754 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:37:52 crc kubenswrapper[4646]: E1203 11:37:52.850712 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:38:04 crc kubenswrapper[4646]: I1203 11:38:04.848416 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:38:04 crc kubenswrapper[4646]: E1203 11:38:04.849219 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:38:18 crc kubenswrapper[4646]: I1203 11:38:18.848637 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:38:18 crc kubenswrapper[4646]: E1203 11:38:18.849981 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:38:30 crc kubenswrapper[4646]: I1203 11:38:30.849230 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:38:30 crc kubenswrapper[4646]: E1203 11:38:30.850174 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:38:44 crc kubenswrapper[4646]: I1203 11:38:44.849963 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:38:44 crc kubenswrapper[4646]: E1203 11:38:44.850757 4646 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pndzt_openshift-machine-config-operator(9695d947-fc43-4625-86af-a1a3e177bb27)\"" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" Dec 03 11:38:59 crc kubenswrapper[4646]: I1203 11:38:59.848146 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" Dec 03 11:39:00 crc kubenswrapper[4646]: I1203 11:39:00.784477 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerStarted","Data":"d6c96feb22a4182604ad8eea667194ee2353c8fca49678c0fb5e3da50a695d32"} Dec 03 11:39:03 crc kubenswrapper[4646]: I1203 11:39:03.809426 4646 generic.go:334] "Generic (PLEG): container finished" podID="6da83524-87b7-401e-ade3-d99a262721e2" containerID="cbcc1dc6529ab7e7f3a79b081cb61dd2cf8f17cdac6a163d545dcfad7b7b0bc4" exitCode=0 Dec 03 11:39:03 crc kubenswrapper[4646]: I1203 11:39:03.809523 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-74vwp/must-gather-fxg6j" event={"ID":"6da83524-87b7-401e-ade3-d99a262721e2","Type":"ContainerDied","Data":"cbcc1dc6529ab7e7f3a79b081cb61dd2cf8f17cdac6a163d545dcfad7b7b0bc4"} Dec 03 11:39:03 crc kubenswrapper[4646]: I1203 11:39:03.811527 4646 scope.go:117] "RemoveContainer" containerID="cbcc1dc6529ab7e7f3a79b081cb61dd2cf8f17cdac6a163d545dcfad7b7b0bc4" Dec 03 11:39:04 crc kubenswrapper[4646]: I1203 11:39:04.678948 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-74vwp_must-gather-fxg6j_6da83524-87b7-401e-ade3-d99a262721e2/gather/0.log" Dec 03 11:39:12 crc kubenswrapper[4646]: I1203 11:39:12.975015 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-74vwp/must-gather-fxg6j"] Dec 03 11:39:12 crc kubenswrapper[4646]: I1203 11:39:12.976074 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-74vwp/must-gather-fxg6j" podUID="6da83524-87b7-401e-ade3-d99a262721e2" containerName="copy" containerID="cri-o://0b94397d08edfaf51ebc7128a4185554039424a85dcbd6c2c709a77ab5e131b6" gracePeriod=2 Dec 03 11:39:13 crc kubenswrapper[4646]: I1203 11:39:13.026949 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-74vwp/must-gather-fxg6j"] Dec 03 11:39:13 crc kubenswrapper[4646]: I1203 11:39:13.549155 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-74vwp_must-gather-fxg6j_6da83524-87b7-401e-ade3-d99a262721e2/copy/0.log" Dec 03 11:39:13 crc kubenswrapper[4646]: I1203 11:39:13.549871 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-74vwp/must-gather-fxg6j" Dec 03 11:39:13 crc kubenswrapper[4646]: I1203 11:39:13.589124 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkjv5\" (UniqueName: \"kubernetes.io/projected/6da83524-87b7-401e-ade3-d99a262721e2-kube-api-access-xkjv5\") pod \"6da83524-87b7-401e-ade3-d99a262721e2\" (UID: \"6da83524-87b7-401e-ade3-d99a262721e2\") " Dec 03 11:39:13 crc kubenswrapper[4646]: I1203 11:39:13.589261 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6da83524-87b7-401e-ade3-d99a262721e2-must-gather-output\") pod \"6da83524-87b7-401e-ade3-d99a262721e2\" (UID: \"6da83524-87b7-401e-ade3-d99a262721e2\") " Dec 03 11:39:13 crc kubenswrapper[4646]: I1203 11:39:13.598704 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6da83524-87b7-401e-ade3-d99a262721e2-kube-api-access-xkjv5" (OuterVolumeSpecName: "kube-api-access-xkjv5") pod "6da83524-87b7-401e-ade3-d99a262721e2" (UID: "6da83524-87b7-401e-ade3-d99a262721e2"). InnerVolumeSpecName "kube-api-access-xkjv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:39:13 crc kubenswrapper[4646]: I1203 11:39:13.691417 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkjv5\" (UniqueName: \"kubernetes.io/projected/6da83524-87b7-401e-ade3-d99a262721e2-kube-api-access-xkjv5\") on node \"crc\" DevicePath \"\"" Dec 03 11:39:13 crc kubenswrapper[4646]: I1203 11:39:13.789477 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6da83524-87b7-401e-ade3-d99a262721e2-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "6da83524-87b7-401e-ade3-d99a262721e2" (UID: "6da83524-87b7-401e-ade3-d99a262721e2"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:39:13 crc kubenswrapper[4646]: I1203 11:39:13.792851 4646 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6da83524-87b7-401e-ade3-d99a262721e2-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 11:39:13 crc kubenswrapper[4646]: I1203 11:39:13.870954 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6da83524-87b7-401e-ade3-d99a262721e2" path="/var/lib/kubelet/pods/6da83524-87b7-401e-ade3-d99a262721e2/volumes" Dec 03 11:39:13 crc kubenswrapper[4646]: I1203 11:39:13.919723 4646 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-74vwp_must-gather-fxg6j_6da83524-87b7-401e-ade3-d99a262721e2/copy/0.log" Dec 03 11:39:13 crc kubenswrapper[4646]: I1203 11:39:13.921634 4646 generic.go:334] "Generic (PLEG): container finished" podID="6da83524-87b7-401e-ade3-d99a262721e2" containerID="0b94397d08edfaf51ebc7128a4185554039424a85dcbd6c2c709a77ab5e131b6" exitCode=143 Dec 03 11:39:13 crc kubenswrapper[4646]: I1203 11:39:13.921717 4646 scope.go:117] "RemoveContainer" containerID="0b94397d08edfaf51ebc7128a4185554039424a85dcbd6c2c709a77ab5e131b6" Dec 03 11:39:13 crc kubenswrapper[4646]: I1203 11:39:13.921900 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-74vwp/must-gather-fxg6j" Dec 03 11:39:14 crc kubenswrapper[4646]: I1203 11:39:14.004206 4646 scope.go:117] "RemoveContainer" containerID="cbcc1dc6529ab7e7f3a79b081cb61dd2cf8f17cdac6a163d545dcfad7b7b0bc4" Dec 03 11:39:14 crc kubenswrapper[4646]: I1203 11:39:14.151508 4646 scope.go:117] "RemoveContainer" containerID="0b94397d08edfaf51ebc7128a4185554039424a85dcbd6c2c709a77ab5e131b6" Dec 03 11:39:14 crc kubenswrapper[4646]: E1203 11:39:14.151991 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b94397d08edfaf51ebc7128a4185554039424a85dcbd6c2c709a77ab5e131b6\": container with ID starting with 0b94397d08edfaf51ebc7128a4185554039424a85dcbd6c2c709a77ab5e131b6 not found: ID does not exist" containerID="0b94397d08edfaf51ebc7128a4185554039424a85dcbd6c2c709a77ab5e131b6" Dec 03 11:39:14 crc kubenswrapper[4646]: I1203 11:39:14.152033 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b94397d08edfaf51ebc7128a4185554039424a85dcbd6c2c709a77ab5e131b6"} err="failed to get container status \"0b94397d08edfaf51ebc7128a4185554039424a85dcbd6c2c709a77ab5e131b6\": rpc error: code = NotFound desc = could not find container \"0b94397d08edfaf51ebc7128a4185554039424a85dcbd6c2c709a77ab5e131b6\": container with ID starting with 0b94397d08edfaf51ebc7128a4185554039424a85dcbd6c2c709a77ab5e131b6 not found: ID does not exist" Dec 03 11:39:14 crc kubenswrapper[4646]: I1203 11:39:14.152061 4646 scope.go:117] "RemoveContainer" containerID="cbcc1dc6529ab7e7f3a79b081cb61dd2cf8f17cdac6a163d545dcfad7b7b0bc4" Dec 03 11:39:14 crc kubenswrapper[4646]: E1203 11:39:14.153488 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbcc1dc6529ab7e7f3a79b081cb61dd2cf8f17cdac6a163d545dcfad7b7b0bc4\": container with ID starting with cbcc1dc6529ab7e7f3a79b081cb61dd2cf8f17cdac6a163d545dcfad7b7b0bc4 not found: ID does not exist" containerID="cbcc1dc6529ab7e7f3a79b081cb61dd2cf8f17cdac6a163d545dcfad7b7b0bc4" Dec 03 11:39:14 crc kubenswrapper[4646]: I1203 11:39:14.153512 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbcc1dc6529ab7e7f3a79b081cb61dd2cf8f17cdac6a163d545dcfad7b7b0bc4"} err="failed to get container status \"cbcc1dc6529ab7e7f3a79b081cb61dd2cf8f17cdac6a163d545dcfad7b7b0bc4\": rpc error: code = NotFound desc = could not find container \"cbcc1dc6529ab7e7f3a79b081cb61dd2cf8f17cdac6a163d545dcfad7b7b0bc4\": container with ID starting with cbcc1dc6529ab7e7f3a79b081cb61dd2cf8f17cdac6a163d545dcfad7b7b0bc4 not found: ID does not exist" Dec 03 11:41:25 crc kubenswrapper[4646]: I1203 11:41:25.964642 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:41:25 crc kubenswrapper[4646]: I1203 11:41:25.965207 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.042135 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-456tb"] Dec 03 11:41:46 crc kubenswrapper[4646]: E1203 11:41:46.043005 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da83524-87b7-401e-ade3-d99a262721e2" containerName="copy" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.043018 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da83524-87b7-401e-ade3-d99a262721e2" containerName="copy" Dec 03 11:41:46 crc kubenswrapper[4646]: E1203 11:41:46.043037 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da83524-87b7-401e-ade3-d99a262721e2" containerName="gather" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.043043 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da83524-87b7-401e-ade3-d99a262721e2" containerName="gather" Dec 03 11:41:46 crc kubenswrapper[4646]: E1203 11:41:46.043065 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38dffc40-f43f-46d4-88ac-8084230ece0f" containerName="extract-content" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.043071 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="38dffc40-f43f-46d4-88ac-8084230ece0f" containerName="extract-content" Dec 03 11:41:46 crc kubenswrapper[4646]: E1203 11:41:46.043083 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38dffc40-f43f-46d4-88ac-8084230ece0f" containerName="extract-utilities" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.043091 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="38dffc40-f43f-46d4-88ac-8084230ece0f" containerName="extract-utilities" Dec 03 11:41:46 crc kubenswrapper[4646]: E1203 11:41:46.043105 4646 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38dffc40-f43f-46d4-88ac-8084230ece0f" containerName="registry-server" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.043112 4646 state_mem.go:107] "Deleted CPUSet assignment" podUID="38dffc40-f43f-46d4-88ac-8084230ece0f" containerName="registry-server" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.043298 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="38dffc40-f43f-46d4-88ac-8084230ece0f" containerName="registry-server" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.043315 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="6da83524-87b7-401e-ade3-d99a262721e2" containerName="gather" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.043323 4646 memory_manager.go:354] "RemoveStaleState removing state" podUID="6da83524-87b7-401e-ade3-d99a262721e2" containerName="copy" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.044735 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.066539 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-456tb"] Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.124734 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-catalog-content\") pod \"redhat-marketplace-456tb\" (UID: \"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6\") " pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.124979 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgd75\" (UniqueName: \"kubernetes.io/projected/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-kube-api-access-xgd75\") pod \"redhat-marketplace-456tb\" (UID: \"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6\") " pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.125115 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-utilities\") pod \"redhat-marketplace-456tb\" (UID: \"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6\") " pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.226972 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgd75\" (UniqueName: \"kubernetes.io/projected/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-kube-api-access-xgd75\") pod \"redhat-marketplace-456tb\" (UID: \"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6\") " pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.227074 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-utilities\") pod \"redhat-marketplace-456tb\" (UID: \"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6\") " pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.227145 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-catalog-content\") pod \"redhat-marketplace-456tb\" (UID: \"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6\") " pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.227643 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-utilities\") pod \"redhat-marketplace-456tb\" (UID: \"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6\") " pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.227660 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-catalog-content\") pod \"redhat-marketplace-456tb\" (UID: \"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6\") " pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.249227 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgd75\" (UniqueName: \"kubernetes.io/projected/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-kube-api-access-xgd75\") pod \"redhat-marketplace-456tb\" (UID: \"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6\") " pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.368522 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:41:46 crc kubenswrapper[4646]: I1203 11:41:46.875074 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-456tb"] Dec 03 11:41:47 crc kubenswrapper[4646]: I1203 11:41:47.253957 4646 generic.go:334] "Generic (PLEG): container finished" podID="38c63f8f-36e4-486b-a6a5-8b172fe3a8f6" containerID="26755b77883f1562cffd53c281d2e324906edd01908a123028fc738f615bce13" exitCode=0 Dec 03 11:41:47 crc kubenswrapper[4646]: I1203 11:41:47.254002 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-456tb" event={"ID":"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6","Type":"ContainerDied","Data":"26755b77883f1562cffd53c281d2e324906edd01908a123028fc738f615bce13"} Dec 03 11:41:47 crc kubenswrapper[4646]: I1203 11:41:47.254027 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-456tb" event={"ID":"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6","Type":"ContainerStarted","Data":"61bdfc13737b0f8ec338587067a2afe5415a73881bcff612f588a404ed522bd5"} Dec 03 11:41:48 crc kubenswrapper[4646]: I1203 11:41:48.266191 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-456tb" event={"ID":"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6","Type":"ContainerStarted","Data":"29532fbc4c88ff2864affa744fcb7a4abff9fd57ac86c3d53907e54c4fb2dde5"} Dec 03 11:41:49 crc kubenswrapper[4646]: I1203 11:41:49.275705 4646 generic.go:334] "Generic (PLEG): container finished" podID="38c63f8f-36e4-486b-a6a5-8b172fe3a8f6" containerID="29532fbc4c88ff2864affa744fcb7a4abff9fd57ac86c3d53907e54c4fb2dde5" exitCode=0 Dec 03 11:41:49 crc kubenswrapper[4646]: I1203 11:41:49.275816 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-456tb" event={"ID":"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6","Type":"ContainerDied","Data":"29532fbc4c88ff2864affa744fcb7a4abff9fd57ac86c3d53907e54c4fb2dde5"} Dec 03 11:41:50 crc kubenswrapper[4646]: I1203 11:41:50.286739 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-456tb" event={"ID":"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6","Type":"ContainerStarted","Data":"8eba9f710240a234f6d0a9a136b2e83c9d4724c5769346e965337a22ca039990"} Dec 03 11:41:50 crc kubenswrapper[4646]: I1203 11:41:50.308801 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-456tb" podStartSLOduration=1.814313636 podStartE2EDuration="4.308780893s" podCreationTimestamp="2025-12-03 11:41:46 +0000 UTC" firstStartedPulling="2025-12-03 11:41:47.255538225 +0000 UTC m=+2883.718594360" lastFinishedPulling="2025-12-03 11:41:49.750005492 +0000 UTC m=+2886.213061617" observedRunningTime="2025-12-03 11:41:50.304542981 +0000 UTC m=+2886.767599116" watchObservedRunningTime="2025-12-03 11:41:50.308780893 +0000 UTC m=+2886.771837028" Dec 03 11:41:55 crc kubenswrapper[4646]: I1203 11:41:55.964418 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:41:55 crc kubenswrapper[4646]: I1203 11:41:55.964913 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:41:56 crc kubenswrapper[4646]: I1203 11:41:56.369514 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:41:56 crc kubenswrapper[4646]: I1203 11:41:56.369626 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:41:56 crc kubenswrapper[4646]: I1203 11:41:56.413498 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:41:57 crc kubenswrapper[4646]: I1203 11:41:57.389926 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:41:57 crc kubenswrapper[4646]: I1203 11:41:57.427952 4646 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qsdtd"] Dec 03 11:41:57 crc kubenswrapper[4646]: I1203 11:41:57.432578 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:41:57 crc kubenswrapper[4646]: I1203 11:41:57.455550 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qsdtd"] Dec 03 11:41:57 crc kubenswrapper[4646]: I1203 11:41:57.475671 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4caadf29-eb5f-424f-a757-7ceccbb51a61-utilities\") pod \"community-operators-qsdtd\" (UID: \"4caadf29-eb5f-424f-a757-7ceccbb51a61\") " pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:41:57 crc kubenswrapper[4646]: I1203 11:41:57.475802 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4caadf29-eb5f-424f-a757-7ceccbb51a61-catalog-content\") pod \"community-operators-qsdtd\" (UID: \"4caadf29-eb5f-424f-a757-7ceccbb51a61\") " pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:41:57 crc kubenswrapper[4646]: I1203 11:41:57.475827 4646 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-785vs\" (UniqueName: \"kubernetes.io/projected/4caadf29-eb5f-424f-a757-7ceccbb51a61-kube-api-access-785vs\") pod \"community-operators-qsdtd\" (UID: \"4caadf29-eb5f-424f-a757-7ceccbb51a61\") " pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:41:57 crc kubenswrapper[4646]: I1203 11:41:57.577989 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4caadf29-eb5f-424f-a757-7ceccbb51a61-catalog-content\") pod \"community-operators-qsdtd\" (UID: \"4caadf29-eb5f-424f-a757-7ceccbb51a61\") " pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:41:57 crc kubenswrapper[4646]: I1203 11:41:57.578034 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-785vs\" (UniqueName: \"kubernetes.io/projected/4caadf29-eb5f-424f-a757-7ceccbb51a61-kube-api-access-785vs\") pod \"community-operators-qsdtd\" (UID: \"4caadf29-eb5f-424f-a757-7ceccbb51a61\") " pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:41:57 crc kubenswrapper[4646]: I1203 11:41:57.578137 4646 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4caadf29-eb5f-424f-a757-7ceccbb51a61-utilities\") pod \"community-operators-qsdtd\" (UID: \"4caadf29-eb5f-424f-a757-7ceccbb51a61\") " pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:41:57 crc kubenswrapper[4646]: I1203 11:41:57.578643 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4caadf29-eb5f-424f-a757-7ceccbb51a61-utilities\") pod \"community-operators-qsdtd\" (UID: \"4caadf29-eb5f-424f-a757-7ceccbb51a61\") " pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:41:57 crc kubenswrapper[4646]: I1203 11:41:57.578649 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4caadf29-eb5f-424f-a757-7ceccbb51a61-catalog-content\") pod \"community-operators-qsdtd\" (UID: \"4caadf29-eb5f-424f-a757-7ceccbb51a61\") " pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:41:57 crc kubenswrapper[4646]: I1203 11:41:57.610398 4646 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-785vs\" (UniqueName: \"kubernetes.io/projected/4caadf29-eb5f-424f-a757-7ceccbb51a61-kube-api-access-785vs\") pod \"community-operators-qsdtd\" (UID: \"4caadf29-eb5f-424f-a757-7ceccbb51a61\") " pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:41:57 crc kubenswrapper[4646]: I1203 11:41:57.755837 4646 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:41:58 crc kubenswrapper[4646]: I1203 11:41:58.083918 4646 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qsdtd"] Dec 03 11:41:58 crc kubenswrapper[4646]: W1203 11:41:58.133965 4646 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4caadf29_eb5f_424f_a757_7ceccbb51a61.slice/crio-3a1a7b7a7cc74995b7289b92d4d22fbd7ddcac550e18604f438a8c3c2dcd52bd WatchSource:0}: Error finding container 3a1a7b7a7cc74995b7289b92d4d22fbd7ddcac550e18604f438a8c3c2dcd52bd: Status 404 returned error can't find the container with id 3a1a7b7a7cc74995b7289b92d4d22fbd7ddcac550e18604f438a8c3c2dcd52bd Dec 03 11:41:58 crc kubenswrapper[4646]: I1203 11:41:58.352587 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qsdtd" event={"ID":"4caadf29-eb5f-424f-a757-7ceccbb51a61","Type":"ContainerStarted","Data":"3a1a7b7a7cc74995b7289b92d4d22fbd7ddcac550e18604f438a8c3c2dcd52bd"} Dec 03 11:41:58 crc kubenswrapper[4646]: I1203 11:41:58.455247 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-456tb"] Dec 03 11:41:59 crc kubenswrapper[4646]: I1203 11:41:59.361814 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qsdtd" event={"ID":"4caadf29-eb5f-424f-a757-7ceccbb51a61","Type":"ContainerStarted","Data":"9bdf941e1f8e6c74fed81fd6f0dca9bc20d813e2ce8e1de6726864734df2aa66"} Dec 03 11:42:00 crc kubenswrapper[4646]: I1203 11:42:00.376631 4646 generic.go:334] "Generic (PLEG): container finished" podID="4caadf29-eb5f-424f-a757-7ceccbb51a61" containerID="9bdf941e1f8e6c74fed81fd6f0dca9bc20d813e2ce8e1de6726864734df2aa66" exitCode=0 Dec 03 11:42:00 crc kubenswrapper[4646]: I1203 11:42:00.377232 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-456tb" podUID="38c63f8f-36e4-486b-a6a5-8b172fe3a8f6" containerName="registry-server" containerID="cri-o://8eba9f710240a234f6d0a9a136b2e83c9d4724c5769346e965337a22ca039990" gracePeriod=2 Dec 03 11:42:00 crc kubenswrapper[4646]: I1203 11:42:00.377399 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qsdtd" event={"ID":"4caadf29-eb5f-424f-a757-7ceccbb51a61","Type":"ContainerDied","Data":"9bdf941e1f8e6c74fed81fd6f0dca9bc20d813e2ce8e1de6726864734df2aa66"} Dec 03 11:42:00 crc kubenswrapper[4646]: I1203 11:42:00.387665 4646 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 11:42:01 crc kubenswrapper[4646]: I1203 11:42:01.389693 4646 generic.go:334] "Generic (PLEG): container finished" podID="38c63f8f-36e4-486b-a6a5-8b172fe3a8f6" containerID="8eba9f710240a234f6d0a9a136b2e83c9d4724c5769346e965337a22ca039990" exitCode=0 Dec 03 11:42:01 crc kubenswrapper[4646]: I1203 11:42:01.389756 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-456tb" event={"ID":"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6","Type":"ContainerDied","Data":"8eba9f710240a234f6d0a9a136b2e83c9d4724c5769346e965337a22ca039990"} Dec 03 11:42:01 crc kubenswrapper[4646]: I1203 11:42:01.752458 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:42:01 crc kubenswrapper[4646]: I1203 11:42:01.870181 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgd75\" (UniqueName: \"kubernetes.io/projected/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-kube-api-access-xgd75\") pod \"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6\" (UID: \"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6\") " Dec 03 11:42:01 crc kubenswrapper[4646]: I1203 11:42:01.870305 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-utilities\") pod \"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6\" (UID: \"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6\") " Dec 03 11:42:01 crc kubenswrapper[4646]: I1203 11:42:01.870407 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-catalog-content\") pod \"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6\" (UID: \"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6\") " Dec 03 11:42:01 crc kubenswrapper[4646]: I1203 11:42:01.874207 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-utilities" (OuterVolumeSpecName: "utilities") pod "38c63f8f-36e4-486b-a6a5-8b172fe3a8f6" (UID: "38c63f8f-36e4-486b-a6a5-8b172fe3a8f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:42:01 crc kubenswrapper[4646]: I1203 11:42:01.876771 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-kube-api-access-xgd75" (OuterVolumeSpecName: "kube-api-access-xgd75") pod "38c63f8f-36e4-486b-a6a5-8b172fe3a8f6" (UID: "38c63f8f-36e4-486b-a6a5-8b172fe3a8f6"). InnerVolumeSpecName "kube-api-access-xgd75". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:42:01 crc kubenswrapper[4646]: I1203 11:42:01.891495 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38c63f8f-36e4-486b-a6a5-8b172fe3a8f6" (UID: "38c63f8f-36e4-486b-a6a5-8b172fe3a8f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:42:01 crc kubenswrapper[4646]: I1203 11:42:01.972407 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 11:42:01 crc kubenswrapper[4646]: I1203 11:42:01.972853 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 11:42:01 crc kubenswrapper[4646]: I1203 11:42:01.972880 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgd75\" (UniqueName: \"kubernetes.io/projected/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6-kube-api-access-xgd75\") on node \"crc\" DevicePath \"\"" Dec 03 11:42:02 crc kubenswrapper[4646]: I1203 11:42:02.399372 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-456tb" event={"ID":"38c63f8f-36e4-486b-a6a5-8b172fe3a8f6","Type":"ContainerDied","Data":"61bdfc13737b0f8ec338587067a2afe5415a73881bcff612f588a404ed522bd5"} Dec 03 11:42:02 crc kubenswrapper[4646]: I1203 11:42:02.399436 4646 scope.go:117] "RemoveContainer" containerID="8eba9f710240a234f6d0a9a136b2e83c9d4724c5769346e965337a22ca039990" Dec 03 11:42:02 crc kubenswrapper[4646]: I1203 11:42:02.399507 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-456tb" Dec 03 11:42:02 crc kubenswrapper[4646]: I1203 11:42:02.431841 4646 scope.go:117] "RemoveContainer" containerID="29532fbc4c88ff2864affa744fcb7a4abff9fd57ac86c3d53907e54c4fb2dde5" Dec 03 11:42:02 crc kubenswrapper[4646]: I1203 11:42:02.433739 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-456tb"] Dec 03 11:42:02 crc kubenswrapper[4646]: I1203 11:42:02.442061 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-456tb"] Dec 03 11:42:02 crc kubenswrapper[4646]: I1203 11:42:02.454885 4646 scope.go:117] "RemoveContainer" containerID="26755b77883f1562cffd53c281d2e324906edd01908a123028fc738f615bce13" Dec 03 11:42:03 crc kubenswrapper[4646]: I1203 11:42:03.879905 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38c63f8f-36e4-486b-a6a5-8b172fe3a8f6" path="/var/lib/kubelet/pods/38c63f8f-36e4-486b-a6a5-8b172fe3a8f6/volumes" Dec 03 11:42:05 crc kubenswrapper[4646]: I1203 11:42:05.434226 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qsdtd" event={"ID":"4caadf29-eb5f-424f-a757-7ceccbb51a61","Type":"ContainerStarted","Data":"01817ad4b46ef8d9004abb7f67e5b5925341bd6229b6db65fd7f4e3f766923a2"} Dec 03 11:42:06 crc kubenswrapper[4646]: I1203 11:42:06.443644 4646 generic.go:334] "Generic (PLEG): container finished" podID="4caadf29-eb5f-424f-a757-7ceccbb51a61" containerID="01817ad4b46ef8d9004abb7f67e5b5925341bd6229b6db65fd7f4e3f766923a2" exitCode=0 Dec 03 11:42:06 crc kubenswrapper[4646]: I1203 11:42:06.443971 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qsdtd" event={"ID":"4caadf29-eb5f-424f-a757-7ceccbb51a61","Type":"ContainerDied","Data":"01817ad4b46ef8d9004abb7f67e5b5925341bd6229b6db65fd7f4e3f766923a2"} Dec 03 11:42:09 crc kubenswrapper[4646]: I1203 11:42:09.473569 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qsdtd" event={"ID":"4caadf29-eb5f-424f-a757-7ceccbb51a61","Type":"ContainerStarted","Data":"0ead75fd99a39e27db60dac9492044f800254f683f1de72b2e097c172a7695d3"} Dec 03 11:42:10 crc kubenswrapper[4646]: I1203 11:42:10.510877 4646 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qsdtd" podStartSLOduration=4.951527935 podStartE2EDuration="13.510860713s" podCreationTimestamp="2025-12-03 11:41:57 +0000 UTC" firstStartedPulling="2025-12-03 11:42:00.385459548 +0000 UTC m=+2896.848515723" lastFinishedPulling="2025-12-03 11:42:08.944792366 +0000 UTC m=+2905.407848501" observedRunningTime="2025-12-03 11:42:10.509511654 +0000 UTC m=+2906.972567789" watchObservedRunningTime="2025-12-03 11:42:10.510860713 +0000 UTC m=+2906.973916848" Dec 03 11:42:17 crc kubenswrapper[4646]: I1203 11:42:17.756991 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:42:17 crc kubenswrapper[4646]: I1203 11:42:17.757914 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:42:17 crc kubenswrapper[4646]: I1203 11:42:17.810535 4646 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:42:18 crc kubenswrapper[4646]: I1203 11:42:18.605061 4646 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:42:18 crc kubenswrapper[4646]: I1203 11:42:18.660777 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qsdtd"] Dec 03 11:42:20 crc kubenswrapper[4646]: I1203 11:42:20.578348 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qsdtd" podUID="4caadf29-eb5f-424f-a757-7ceccbb51a61" containerName="registry-server" containerID="cri-o://0ead75fd99a39e27db60dac9492044f800254f683f1de72b2e097c172a7695d3" gracePeriod=2 Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.002107 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.097732 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-785vs\" (UniqueName: \"kubernetes.io/projected/4caadf29-eb5f-424f-a757-7ceccbb51a61-kube-api-access-785vs\") pod \"4caadf29-eb5f-424f-a757-7ceccbb51a61\" (UID: \"4caadf29-eb5f-424f-a757-7ceccbb51a61\") " Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.097949 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4caadf29-eb5f-424f-a757-7ceccbb51a61-utilities\") pod \"4caadf29-eb5f-424f-a757-7ceccbb51a61\" (UID: \"4caadf29-eb5f-424f-a757-7ceccbb51a61\") " Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.098012 4646 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4caadf29-eb5f-424f-a757-7ceccbb51a61-catalog-content\") pod \"4caadf29-eb5f-424f-a757-7ceccbb51a61\" (UID: \"4caadf29-eb5f-424f-a757-7ceccbb51a61\") " Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.099006 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4caadf29-eb5f-424f-a757-7ceccbb51a61-utilities" (OuterVolumeSpecName: "utilities") pod "4caadf29-eb5f-424f-a757-7ceccbb51a61" (UID: "4caadf29-eb5f-424f-a757-7ceccbb51a61"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.111855 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4caadf29-eb5f-424f-a757-7ceccbb51a61-kube-api-access-785vs" (OuterVolumeSpecName: "kube-api-access-785vs") pod "4caadf29-eb5f-424f-a757-7ceccbb51a61" (UID: "4caadf29-eb5f-424f-a757-7ceccbb51a61"). InnerVolumeSpecName "kube-api-access-785vs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.150577 4646 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4caadf29-eb5f-424f-a757-7ceccbb51a61-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4caadf29-eb5f-424f-a757-7ceccbb51a61" (UID: "4caadf29-eb5f-424f-a757-7ceccbb51a61"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.199960 4646 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-785vs\" (UniqueName: \"kubernetes.io/projected/4caadf29-eb5f-424f-a757-7ceccbb51a61-kube-api-access-785vs\") on node \"crc\" DevicePath \"\"" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.200000 4646 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4caadf29-eb5f-424f-a757-7ceccbb51a61-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.200013 4646 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4caadf29-eb5f-424f-a757-7ceccbb51a61-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.588073 4646 generic.go:334] "Generic (PLEG): container finished" podID="4caadf29-eb5f-424f-a757-7ceccbb51a61" containerID="0ead75fd99a39e27db60dac9492044f800254f683f1de72b2e097c172a7695d3" exitCode=0 Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.588144 4646 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qsdtd" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.590641 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qsdtd" event={"ID":"4caadf29-eb5f-424f-a757-7ceccbb51a61","Type":"ContainerDied","Data":"0ead75fd99a39e27db60dac9492044f800254f683f1de72b2e097c172a7695d3"} Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.590748 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qsdtd" event={"ID":"4caadf29-eb5f-424f-a757-7ceccbb51a61","Type":"ContainerDied","Data":"3a1a7b7a7cc74995b7289b92d4d22fbd7ddcac550e18604f438a8c3c2dcd52bd"} Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.590834 4646 scope.go:117] "RemoveContainer" containerID="0ead75fd99a39e27db60dac9492044f800254f683f1de72b2e097c172a7695d3" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.630882 4646 scope.go:117] "RemoveContainer" containerID="01817ad4b46ef8d9004abb7f67e5b5925341bd6229b6db65fd7f4e3f766923a2" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.630939 4646 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qsdtd"] Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.644778 4646 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qsdtd"] Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.651031 4646 scope.go:117] "RemoveContainer" containerID="9bdf941e1f8e6c74fed81fd6f0dca9bc20d813e2ce8e1de6726864734df2aa66" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.696607 4646 scope.go:117] "RemoveContainer" containerID="0ead75fd99a39e27db60dac9492044f800254f683f1de72b2e097c172a7695d3" Dec 03 11:42:21 crc kubenswrapper[4646]: E1203 11:42:21.697870 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ead75fd99a39e27db60dac9492044f800254f683f1de72b2e097c172a7695d3\": container with ID starting with 0ead75fd99a39e27db60dac9492044f800254f683f1de72b2e097c172a7695d3 not found: ID does not exist" containerID="0ead75fd99a39e27db60dac9492044f800254f683f1de72b2e097c172a7695d3" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.697911 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ead75fd99a39e27db60dac9492044f800254f683f1de72b2e097c172a7695d3"} err="failed to get container status \"0ead75fd99a39e27db60dac9492044f800254f683f1de72b2e097c172a7695d3\": rpc error: code = NotFound desc = could not find container \"0ead75fd99a39e27db60dac9492044f800254f683f1de72b2e097c172a7695d3\": container with ID starting with 0ead75fd99a39e27db60dac9492044f800254f683f1de72b2e097c172a7695d3 not found: ID does not exist" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.697933 4646 scope.go:117] "RemoveContainer" containerID="01817ad4b46ef8d9004abb7f67e5b5925341bd6229b6db65fd7f4e3f766923a2" Dec 03 11:42:21 crc kubenswrapper[4646]: E1203 11:42:21.698296 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01817ad4b46ef8d9004abb7f67e5b5925341bd6229b6db65fd7f4e3f766923a2\": container with ID starting with 01817ad4b46ef8d9004abb7f67e5b5925341bd6229b6db65fd7f4e3f766923a2 not found: ID does not exist" containerID="01817ad4b46ef8d9004abb7f67e5b5925341bd6229b6db65fd7f4e3f766923a2" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.698323 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01817ad4b46ef8d9004abb7f67e5b5925341bd6229b6db65fd7f4e3f766923a2"} err="failed to get container status \"01817ad4b46ef8d9004abb7f67e5b5925341bd6229b6db65fd7f4e3f766923a2\": rpc error: code = NotFound desc = could not find container \"01817ad4b46ef8d9004abb7f67e5b5925341bd6229b6db65fd7f4e3f766923a2\": container with ID starting with 01817ad4b46ef8d9004abb7f67e5b5925341bd6229b6db65fd7f4e3f766923a2 not found: ID does not exist" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.698353 4646 scope.go:117] "RemoveContainer" containerID="9bdf941e1f8e6c74fed81fd6f0dca9bc20d813e2ce8e1de6726864734df2aa66" Dec 03 11:42:21 crc kubenswrapper[4646]: E1203 11:42:21.698692 4646 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bdf941e1f8e6c74fed81fd6f0dca9bc20d813e2ce8e1de6726864734df2aa66\": container with ID starting with 9bdf941e1f8e6c74fed81fd6f0dca9bc20d813e2ce8e1de6726864734df2aa66 not found: ID does not exist" containerID="9bdf941e1f8e6c74fed81fd6f0dca9bc20d813e2ce8e1de6726864734df2aa66" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.698717 4646 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bdf941e1f8e6c74fed81fd6f0dca9bc20d813e2ce8e1de6726864734df2aa66"} err="failed to get container status \"9bdf941e1f8e6c74fed81fd6f0dca9bc20d813e2ce8e1de6726864734df2aa66\": rpc error: code = NotFound desc = could not find container \"9bdf941e1f8e6c74fed81fd6f0dca9bc20d813e2ce8e1de6726864734df2aa66\": container with ID starting with 9bdf941e1f8e6c74fed81fd6f0dca9bc20d813e2ce8e1de6726864734df2aa66 not found: ID does not exist" Dec 03 11:42:21 crc kubenswrapper[4646]: I1203 11:42:21.861475 4646 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4caadf29-eb5f-424f-a757-7ceccbb51a61" path="/var/lib/kubelet/pods/4caadf29-eb5f-424f-a757-7ceccbb51a61/volumes" Dec 03 11:42:25 crc kubenswrapper[4646]: I1203 11:42:25.964600 4646 patch_prober.go:28] interesting pod/machine-config-daemon-pndzt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 11:42:25 crc kubenswrapper[4646]: I1203 11:42:25.967685 4646 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 11:42:25 crc kubenswrapper[4646]: I1203 11:42:25.967930 4646 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" Dec 03 11:42:25 crc kubenswrapper[4646]: I1203 11:42:25.969478 4646 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d6c96feb22a4182604ad8eea667194ee2353c8fca49678c0fb5e3da50a695d32"} pod="openshift-machine-config-operator/machine-config-daemon-pndzt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 11:42:25 crc kubenswrapper[4646]: I1203 11:42:25.969673 4646 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" podUID="9695d947-fc43-4625-86af-a1a3e177bb27" containerName="machine-config-daemon" containerID="cri-o://d6c96feb22a4182604ad8eea667194ee2353c8fca49678c0fb5e3da50a695d32" gracePeriod=600 Dec 03 11:42:26 crc kubenswrapper[4646]: I1203 11:42:26.652794 4646 generic.go:334] "Generic (PLEG): container finished" podID="9695d947-fc43-4625-86af-a1a3e177bb27" containerID="d6c96feb22a4182604ad8eea667194ee2353c8fca49678c0fb5e3da50a695d32" exitCode=0 Dec 03 11:42:26 crc kubenswrapper[4646]: I1203 11:42:26.652866 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerDied","Data":"d6c96feb22a4182604ad8eea667194ee2353c8fca49678c0fb5e3da50a695d32"} Dec 03 11:42:26 crc kubenswrapper[4646]: I1203 11:42:26.653600 4646 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pndzt" event={"ID":"9695d947-fc43-4625-86af-a1a3e177bb27","Type":"ContainerStarted","Data":"3f73544f7f259f6422de4e59acaca480360c90afa3473fd433936788271d18ba"} Dec 03 11:42:26 crc kubenswrapper[4646]: I1203 11:42:26.653672 4646 scope.go:117] "RemoveContainer" containerID="19aea92baf9726284615749ddaf0302cb265ee3a476892205db63e1792e67d29" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114021143024435 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114021144017353 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114012656016506 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114012657015457 5ustar corecore